var/home/core/zuul-output/0000755000175000017500000000000015115003052014515 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015115016106015465 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005257201515115016100017672 0ustar rootrootDec 06 10:36:34 crc systemd[1]: Starting Kubernetes Kubelet... Dec 06 10:36:34 crc restorecon[4578]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:34 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 06 10:36:35 crc restorecon[4578]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 06 10:36:35 crc kubenswrapper[4678]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 10:36:35 crc kubenswrapper[4678]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 06 10:36:35 crc kubenswrapper[4678]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 10:36:35 crc kubenswrapper[4678]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 10:36:35 crc kubenswrapper[4678]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 06 10:36:35 crc kubenswrapper[4678]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.279800 4678 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282613 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282630 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282635 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282640 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282646 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282652 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282658 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282663 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282668 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282673 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282679 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282684 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282689 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282694 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282698 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282703 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282707 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282712 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282716 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282721 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282725 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282737 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282741 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282746 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282750 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282755 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282760 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282764 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282768 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282776 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282783 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282789 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282794 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282798 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282803 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282808 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282812 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282817 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282822 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282827 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282831 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282836 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282841 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282847 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282853 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282859 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282867 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282873 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282878 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282883 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282887 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282892 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282896 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282901 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282908 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282913 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282917 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282924 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282929 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282934 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282938 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282943 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282948 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282953 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282957 4678 feature_gate.go:330] unrecognized feature gate: Example Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282961 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282965 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282970 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282975 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282979 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.282984 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283097 4678 flags.go:64] FLAG: --address="0.0.0.0" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283111 4678 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283122 4678 flags.go:64] FLAG: --anonymous-auth="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283130 4678 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283137 4678 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283142 4678 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283150 4678 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283158 4678 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283165 4678 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283170 4678 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283176 4678 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283182 4678 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283188 4678 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283194 4678 flags.go:64] FLAG: --cgroup-root="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283199 4678 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283205 4678 flags.go:64] FLAG: --client-ca-file="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283211 4678 flags.go:64] FLAG: --cloud-config="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283217 4678 flags.go:64] FLAG: --cloud-provider="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283223 4678 flags.go:64] FLAG: --cluster-dns="[]" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283230 4678 flags.go:64] FLAG: --cluster-domain="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283236 4678 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283241 4678 flags.go:64] FLAG: --config-dir="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283247 4678 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283254 4678 flags.go:64] FLAG: --container-log-max-files="5" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283262 4678 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283268 4678 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283275 4678 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283281 4678 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283286 4678 flags.go:64] FLAG: --contention-profiling="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283292 4678 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283297 4678 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283303 4678 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283310 4678 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283317 4678 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283321 4678 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283327 4678 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283331 4678 flags.go:64] FLAG: --enable-load-reader="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283336 4678 flags.go:64] FLAG: --enable-server="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283341 4678 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283348 4678 flags.go:64] FLAG: --event-burst="100" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283353 4678 flags.go:64] FLAG: --event-qps="50" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283358 4678 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283364 4678 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283369 4678 flags.go:64] FLAG: --eviction-hard="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283380 4678 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283385 4678 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283390 4678 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283397 4678 flags.go:64] FLAG: --eviction-soft="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283402 4678 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283407 4678 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283412 4678 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283417 4678 flags.go:64] FLAG: --experimental-mounter-path="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283422 4678 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283427 4678 flags.go:64] FLAG: --fail-swap-on="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283431 4678 flags.go:64] FLAG: --feature-gates="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283438 4678 flags.go:64] FLAG: --file-check-frequency="20s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283442 4678 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283447 4678 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283452 4678 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283457 4678 flags.go:64] FLAG: --healthz-port="10248" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283462 4678 flags.go:64] FLAG: --help="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283467 4678 flags.go:64] FLAG: --hostname-override="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283472 4678 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283477 4678 flags.go:64] FLAG: --http-check-frequency="20s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283500 4678 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283506 4678 flags.go:64] FLAG: --image-credential-provider-config="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283511 4678 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283516 4678 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283523 4678 flags.go:64] FLAG: --image-service-endpoint="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283528 4678 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283534 4678 flags.go:64] FLAG: --kube-api-burst="100" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283539 4678 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283545 4678 flags.go:64] FLAG: --kube-api-qps="50" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283550 4678 flags.go:64] FLAG: --kube-reserved="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283556 4678 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283561 4678 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283566 4678 flags.go:64] FLAG: --kubelet-cgroups="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283571 4678 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283577 4678 flags.go:64] FLAG: --lock-file="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283582 4678 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283589 4678 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283596 4678 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283605 4678 flags.go:64] FLAG: --log-json-split-stream="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283610 4678 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283615 4678 flags.go:64] FLAG: --log-text-split-stream="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283621 4678 flags.go:64] FLAG: --logging-format="text" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283626 4678 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283632 4678 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283637 4678 flags.go:64] FLAG: --manifest-url="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283642 4678 flags.go:64] FLAG: --manifest-url-header="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283650 4678 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283655 4678 flags.go:64] FLAG: --max-open-files="1000000" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283662 4678 flags.go:64] FLAG: --max-pods="110" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283667 4678 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283672 4678 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283678 4678 flags.go:64] FLAG: --memory-manager-policy="None" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283682 4678 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283687 4678 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283692 4678 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283697 4678 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283709 4678 flags.go:64] FLAG: --node-status-max-images="50" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283714 4678 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283719 4678 flags.go:64] FLAG: --oom-score-adj="-999" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283725 4678 flags.go:64] FLAG: --pod-cidr="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283731 4678 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283740 4678 flags.go:64] FLAG: --pod-manifest-path="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283745 4678 flags.go:64] FLAG: --pod-max-pids="-1" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283750 4678 flags.go:64] FLAG: --pods-per-core="0" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283755 4678 flags.go:64] FLAG: --port="10250" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283760 4678 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283765 4678 flags.go:64] FLAG: --provider-id="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283770 4678 flags.go:64] FLAG: --qos-reserved="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283775 4678 flags.go:64] FLAG: --read-only-port="10255" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283784 4678 flags.go:64] FLAG: --register-node="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283789 4678 flags.go:64] FLAG: --register-schedulable="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283794 4678 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283804 4678 flags.go:64] FLAG: --registry-burst="10" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283810 4678 flags.go:64] FLAG: --registry-qps="5" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283815 4678 flags.go:64] FLAG: --reserved-cpus="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283821 4678 flags.go:64] FLAG: --reserved-memory="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283828 4678 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283834 4678 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283839 4678 flags.go:64] FLAG: --rotate-certificates="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283844 4678 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283849 4678 flags.go:64] FLAG: --runonce="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283854 4678 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283860 4678 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283866 4678 flags.go:64] FLAG: --seccomp-default="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283871 4678 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283876 4678 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283882 4678 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283888 4678 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283894 4678 flags.go:64] FLAG: --storage-driver-password="root" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283899 4678 flags.go:64] FLAG: --storage-driver-secure="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283905 4678 flags.go:64] FLAG: --storage-driver-table="stats" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283910 4678 flags.go:64] FLAG: --storage-driver-user="root" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283916 4678 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283922 4678 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283929 4678 flags.go:64] FLAG: --system-cgroups="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283934 4678 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283944 4678 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283949 4678 flags.go:64] FLAG: --tls-cert-file="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283954 4678 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283962 4678 flags.go:64] FLAG: --tls-min-version="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283967 4678 flags.go:64] FLAG: --tls-private-key-file="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283975 4678 flags.go:64] FLAG: --topology-manager-policy="none" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283981 4678 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283987 4678 flags.go:64] FLAG: --topology-manager-scope="container" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283992 4678 flags.go:64] FLAG: --v="2" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.283999 4678 flags.go:64] FLAG: --version="false" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.284006 4678 flags.go:64] FLAG: --vmodule="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.284013 4678 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.284018 4678 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284163 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284171 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284177 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284182 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284187 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284194 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284199 4678 feature_gate.go:330] unrecognized feature gate: Example Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284204 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284209 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284213 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284218 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284223 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284227 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284232 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284236 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284241 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284246 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284250 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284255 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284260 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284265 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284269 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284274 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284280 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284289 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284294 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284299 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284303 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284308 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284312 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284317 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284321 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284326 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284331 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284335 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284340 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284345 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284350 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284354 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284359 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284365 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284372 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284377 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284382 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284386 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284391 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284398 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284404 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284408 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284413 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284418 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284424 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284429 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284434 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284439 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284443 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284451 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284456 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284460 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284467 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284472 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284478 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284501 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284507 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284511 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284516 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284522 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284528 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284534 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284539 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.284544 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.284729 4678 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.295670 4678 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.295743 4678 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295873 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295896 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295906 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295917 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295929 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295941 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295950 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295959 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295967 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295976 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295984 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.295992 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296000 4678 feature_gate.go:330] unrecognized feature gate: Example Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296009 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296018 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296028 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296037 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296045 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296053 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296061 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296069 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296080 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296090 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296099 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296108 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296117 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296126 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296134 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296143 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296152 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296160 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296168 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296176 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296184 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296194 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296202 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296210 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296218 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296226 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296235 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296247 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296262 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296309 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296318 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296325 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296333 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296344 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296352 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296360 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296368 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296375 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296383 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296391 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296402 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296411 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296420 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296429 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296438 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296447 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296455 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296464 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296473 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296481 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296507 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296516 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296529 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296538 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296546 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296554 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296562 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296571 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.296585 4678 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296832 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296849 4678 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296858 4678 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296866 4678 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296876 4678 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296884 4678 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296892 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296902 4678 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296910 4678 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296920 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296928 4678 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296936 4678 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296944 4678 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296951 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296959 4678 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296967 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296975 4678 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296983 4678 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296991 4678 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.296999 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297007 4678 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297015 4678 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297025 4678 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297035 4678 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297043 4678 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297053 4678 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297063 4678 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297074 4678 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297083 4678 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297091 4678 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297100 4678 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297108 4678 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297115 4678 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297123 4678 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297131 4678 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297139 4678 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297147 4678 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297155 4678 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297162 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297170 4678 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297178 4678 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297186 4678 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297194 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297203 4678 feature_gate.go:330] unrecognized feature gate: Example Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297212 4678 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297220 4678 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297228 4678 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297236 4678 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297244 4678 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297253 4678 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297261 4678 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297268 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297277 4678 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297284 4678 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297292 4678 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297299 4678 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297307 4678 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297315 4678 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297323 4678 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297334 4678 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297343 4678 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297352 4678 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297360 4678 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297369 4678 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297377 4678 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297385 4678 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297393 4678 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297400 4678 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297411 4678 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297420 4678 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.297430 4678 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.297443 4678 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.298084 4678 server.go:940] "Client rotation is on, will bootstrap in background" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.303189 4678 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.303368 4678 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.304298 4678 server.go:997] "Starting client certificate rotation" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.304362 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.304789 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-13 22:42:52.719398675 +0000 UTC Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.304884 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.312035 4678 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.317035 4678 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.319193 4678 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.327006 4678 log.go:25] "Validated CRI v1 runtime API" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.359607 4678 log.go:25] "Validated CRI v1 image API" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.362389 4678 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.365893 4678 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-06-10-31-06-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.366014 4678 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.386279 4678 manager.go:217] Machine: {Timestamp:2025-12-06 10:36:35.384997265 +0000 UTC m=+0.228428744 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199476736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:05fdc3ed-7324-40a6-b64a-9e2883945f0d BootID:d036a3e2-4349-4778-85ff-e71cad7388b7 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:8d:bc:bc Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:8d:bc:bc Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:4b:53:ea Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e7:3f:ce Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e6:f4:e8 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:5d:a2:68 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:06:fa:8e:73:f2:43 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:b2:d7:84:5e:2c:be Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199476736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.386603 4678 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.386762 4678 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.387369 4678 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.387617 4678 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.387669 4678 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.387935 4678 topology_manager.go:138] "Creating topology manager with none policy" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.387953 4678 container_manager_linux.go:303] "Creating device plugin manager" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.388250 4678 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.388301 4678 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.388534 4678 state_mem.go:36] "Initialized new in-memory state store" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.388643 4678 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.389981 4678 kubelet.go:418] "Attempting to sync node with API server" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.390016 4678 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.390054 4678 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.390078 4678 kubelet.go:324] "Adding apiserver pod source" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.390094 4678 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.391704 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.391819 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.391956 4678 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.392141 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.392341 4678 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.392602 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.393983 4678 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.397907 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398380 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398405 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398415 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398431 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398442 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398452 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398465 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398518 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398529 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398544 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398554 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.398581 4678 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.399544 4678 server.go:1280] "Started kubelet" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.400323 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.400656 4678 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.400621 4678 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 06 10:36:35 crc systemd[1]: Started Kubernetes Kubelet. Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.402366 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.402405 4678 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.402512 4678 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.402950 4678 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.402968 4678 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.403074 4678 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.402601 4678 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.198:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e99f83b2bbb8c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 10:36:35.39947406 +0000 UTC m=+0.242905499,LastTimestamp:2025-12-06 10:36:35.39947406 +0000 UTC m=+0.242905499,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.403081 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 02:19:00.389021531 +0000 UTC Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.403223 4678 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.404237 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="200ms" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.404026 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.404421 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.406326 4678 server.go:460] "Adding debug handlers to kubelet server" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.410753 4678 factory.go:55] Registering systemd factory Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411141 4678 factory.go:221] Registration of the systemd container factory successfully Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.410835 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411400 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411468 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411556 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411585 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411608 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411634 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411659 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411697 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411729 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411761 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411794 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411822 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411856 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411880 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411902 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411930 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.411960 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412022 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412046 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412067 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412090 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412112 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412133 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412155 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412180 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412214 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412242 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412264 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412286 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412307 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412340 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412363 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412384 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412408 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412431 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412452 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412473 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412532 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412565 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412592 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412618 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412640 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412662 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412683 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412711 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412737 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412759 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412781 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412805 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412828 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412849 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412878 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412901 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412929 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412959 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.412989 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413020 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413041 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413061 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413085 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413104 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413124 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413145 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413165 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413184 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413207 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413226 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413245 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413266 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.413289 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.414222 4678 factory.go:153] Registering CRI-O factory Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.414249 4678 factory.go:221] Registration of the crio container factory successfully Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.414340 4678 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.414366 4678 factory.go:103] Registering Raw factory Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.414386 4678 manager.go:1196] Started watching for new ooms in manager Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.414890 4678 manager.go:319] Starting recovery of all containers Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423480 4678 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423570 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423598 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423625 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423643 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423664 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423685 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423702 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423725 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423742 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423762 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423778 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423795 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423823 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423844 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423869 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423897 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423915 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423944 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423960 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423978 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.423999 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424020 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424041 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424060 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424077 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424098 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424114 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424134 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424153 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424171 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424194 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424209 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424237 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424275 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424306 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424334 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.424932 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.425042 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.425244 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.425325 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.427145 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.428108 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.428205 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.428711 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.428805 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.428889 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.428973 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.429054 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.429147 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.430095 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.430667 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.431843 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437063 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437107 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437138 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437160 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437176 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437204 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437223 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437245 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437263 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437279 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437301 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437318 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437340 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.431442 4678 manager.go:324] Recovery completed Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437390 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437409 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437431 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437450 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437473 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437574 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437592 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437613 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437632 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437662 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437677 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437692 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437711 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437726 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437741 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437762 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437776 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437796 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437812 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437832 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437857 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437873 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437890 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437905 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437921 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437940 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437956 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437975 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.437990 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438007 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438028 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438045 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438063 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438079 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438114 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438133 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438148 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438166 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438181 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438196 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438216 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438232 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438250 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438265 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438280 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438299 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438317 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438337 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438356 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438371 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438389 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438405 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438420 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438438 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438452 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438471 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438507 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438523 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438543 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438560 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438578 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438595 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438611 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438631 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438650 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438670 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438685 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438701 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438720 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438735 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438769 4678 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438783 4678 reconstruct.go:97] "Volume reconstruction finished" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.438793 4678 reconciler.go:26] "Reconciler: start to sync state" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.459636 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.462658 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.462692 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.462705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.464244 4678 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.464271 4678 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.464301 4678 state_mem.go:36] "Initialized new in-memory state store" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.469404 4678 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.474608 4678 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.474650 4678 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.474681 4678 kubelet.go:2335] "Starting kubelet main sync loop" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.474736 4678 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.475717 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.475780 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.480583 4678 policy_none.go:49] "None policy: Start" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.481448 4678 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.481503 4678 state_mem.go:35] "Initializing new in-memory state store" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.503363 4678 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.531466 4678 manager.go:334] "Starting Device Plugin manager" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.531552 4678 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.531573 4678 server.go:79] "Starting device plugin registration server" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.532140 4678 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.532162 4678 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.533404 4678 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.533531 4678 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.533549 4678 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.545524 4678 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.575460 4678 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.575642 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.577053 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.577112 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.577124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.577338 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.577527 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.577588 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.578640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.578694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.578705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.578785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.578811 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.578825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.578898 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.579218 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.579267 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.579858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.579895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.579965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.580141 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.580197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.580231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.580252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.580256 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.580367 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581111 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581232 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581424 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581582 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.581641 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582601 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582649 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582657 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.582669 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.583332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.583363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.583375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.604810 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="400ms" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.632434 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.634526 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.634591 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.634605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.634645 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.635476 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645578 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645623 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645728 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645774 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645847 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645898 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645932 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.645993 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.646094 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.646159 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.646194 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.646226 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.646255 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.646279 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.646300 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747436 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747543 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747577 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747614 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747649 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747682 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747715 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747759 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747800 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747803 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747913 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747871 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747933 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747822 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747832 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747917 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748074 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747973 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748101 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748125 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748133 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.747821 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748144 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748188 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748225 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748254 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748265 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.748402 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.836386 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.838033 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.838069 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.838082 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.838108 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 10:36:35 crc kubenswrapper[4678]: E1206 10:36:35.838586 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.902393 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.911167 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.929255 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.943224 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-978fd6f05da8d7d6b66a94b5e5094596f95365b25a2227673509503e7cf4b3bd WatchSource:0}: Error finding container 978fd6f05da8d7d6b66a94b5e5094596f95365b25a2227673509503e7cf4b3bd: Status 404 returned error can't find the container with id 978fd6f05da8d7d6b66a94b5e5094596f95365b25a2227673509503e7cf4b3bd Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.943995 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ff3d2e2901e59a4e8e2422419f25db5c78e252ee09778ead792268b9c6f6ac00 WatchSource:0}: Error finding container ff3d2e2901e59a4e8e2422419f25db5c78e252ee09778ead792268b9c6f6ac00: Status 404 returned error can't find the container with id ff3d2e2901e59a4e8e2422419f25db5c78e252ee09778ead792268b9c6f6ac00 Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.950686 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.954261 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-93487e0f6902cdebaa18d15f7be3e4e0b4755d51154706a84737269058e3c1e3 WatchSource:0}: Error finding container 93487e0f6902cdebaa18d15f7be3e4e0b4755d51154706a84737269058e3c1e3: Status 404 returned error can't find the container with id 93487e0f6902cdebaa18d15f7be3e4e0b4755d51154706a84737269058e3c1e3 Dec 06 10:36:35 crc kubenswrapper[4678]: I1206 10:36:35.959169 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:35 crc kubenswrapper[4678]: W1206 10:36:35.983110 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ee8247194dcb6bd5efdd6fe14ab25745bd880be4516b45a2c30273fb77a81a7c WatchSource:0}: Error finding container ee8247194dcb6bd5efdd6fe14ab25745bd880be4516b45a2c30273fb77a81a7c: Status 404 returned error can't find the container with id ee8247194dcb6bd5efdd6fe14ab25745bd880be4516b45a2c30273fb77a81a7c Dec 06 10:36:36 crc kubenswrapper[4678]: E1206 10:36:36.005901 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="800ms" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.239119 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.240847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.240897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.240913 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.240950 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 10:36:36 crc kubenswrapper[4678]: E1206 10:36:36.241621 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.402806 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.404098 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 22:05:53.289270611 +0000 UTC Dec 06 10:36:36 crc kubenswrapper[4678]: W1206 10:36:36.437913 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:36 crc kubenswrapper[4678]: E1206 10:36:36.438036 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.481350 4678 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38" exitCode=0 Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.481461 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.481879 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"978fd6f05da8d7d6b66a94b5e5094596f95365b25a2227673509503e7cf4b3bd"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.482069 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.483645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.483717 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.483736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.484688 4678 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="69956e54a4d82ac7a8c7d6e7bce9bb238e0d4d13d857ed8de5aa2dbf6980ff58" exitCode=0 Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.484756 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"69956e54a4d82ac7a8c7d6e7bce9bb238e0d4d13d857ed8de5aa2dbf6980ff58"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.484813 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ff3d2e2901e59a4e8e2422419f25db5c78e252ee09778ead792268b9c6f6ac00"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.484922 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.486402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.486472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.486542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.486762 4678 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539" exitCode=0 Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.486852 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.486918 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ee8247194dcb6bd5efdd6fe14ab25745bd880be4516b45a2c30273fb77a81a7c"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.487020 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.487956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.487989 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.488004 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.491555 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.491614 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"83862366630bb3dfee4348c2cd899114eae52d70f6f50a7b9b4bb25f798afdbd"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.499217 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222" exitCode=0 Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.499317 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.499404 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"93487e0f6902cdebaa18d15f7be3e4e0b4755d51154706a84737269058e3c1e3"} Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.499914 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.505401 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.505457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.505477 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.513589 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.514953 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.515002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:36 crc kubenswrapper[4678]: I1206 10:36:36.515019 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:36 crc kubenswrapper[4678]: W1206 10:36:36.570272 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:36 crc kubenswrapper[4678]: E1206 10:36:36.570369 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:36 crc kubenswrapper[4678]: W1206 10:36:36.782753 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:36 crc kubenswrapper[4678]: E1206 10:36:36.782985 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:36 crc kubenswrapper[4678]: E1206 10:36:36.807218 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="1.6s" Dec 06 10:36:36 crc kubenswrapper[4678]: W1206 10:36:36.817611 4678 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:36 crc kubenswrapper[4678]: E1206 10:36:36.817707 4678 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.042021 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.043887 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.043920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.043930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.044000 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 10:36:37 crc kubenswrapper[4678]: E1206 10:36:37.044598 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.198:6443: connect: connection refused" node="crc" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.401780 4678 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.198:6443: connect: connection refused Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.404853 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 13:49:23.017131349 +0000 UTC Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.404922 4678 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 795h12m45.612211959s for next certificate rotation Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.434034 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 10:36:37 crc kubenswrapper[4678]: E1206 10:36:37.435578 4678 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.198:6443: connect: connection refused" logger="UnhandledError" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.504845 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.504895 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.504905 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.505032 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.505690 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.505717 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.505726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.508308 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.508298 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.508459 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.508481 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.508930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.508952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.508962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.513607 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.513658 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.513669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.513678 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.517002 4678 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40" exitCode=0 Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.517105 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.517261 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.518246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.518304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.518316 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.525774 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a5b15300e06c9969cb618aa54a705101e37282d0992fd20ed8bb251bee3ef297"} Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.525896 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.527096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.527127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:37 crc kubenswrapper[4678]: I1206 10:36:37.527141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.534770 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d"} Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.534895 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.536546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.536601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.536621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.538715 4678 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185" exitCode=0 Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.538749 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185"} Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.539000 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.539143 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.540340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.540416 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.540446 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.540426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.540615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.540629 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.645050 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.646825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.646918 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.646941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:38 crc kubenswrapper[4678]: I1206 10:36:38.646979 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.519098 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.547224 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7"} Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.547293 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9"} Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.547320 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2"} Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.547331 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.547336 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e"} Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.547548 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.548351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.548393 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:39 crc kubenswrapper[4678]: I1206 10:36:39.548413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.552334 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.556007 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688"} Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.556098 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.556186 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.557413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.557472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.557533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.558748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.558778 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:40 crc kubenswrapper[4678]: I1206 10:36:40.558789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.519965 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.559529 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.559533 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.560809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.560867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.560889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.560884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.560927 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:41 crc kubenswrapper[4678]: I1206 10:36:41.560935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.122600 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.122840 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.124294 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.124367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.124383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.623013 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.623896 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.625726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.625774 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:42 crc kubenswrapper[4678]: I1206 10:36:42.625793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:43 crc kubenswrapper[4678]: I1206 10:36:43.933645 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:43 crc kubenswrapper[4678]: I1206 10:36:43.933945 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:43 crc kubenswrapper[4678]: I1206 10:36:43.935820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:43 crc kubenswrapper[4678]: I1206 10:36:43.935896 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:43 crc kubenswrapper[4678]: I1206 10:36:43.936093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:43 crc kubenswrapper[4678]: I1206 10:36:43.941435 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:44 crc kubenswrapper[4678]: I1206 10:36:44.569214 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:44 crc kubenswrapper[4678]: I1206 10:36:44.570839 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:44 crc kubenswrapper[4678]: I1206 10:36:44.570875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:44 crc kubenswrapper[4678]: I1206 10:36:44.570889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:44 crc kubenswrapper[4678]: I1206 10:36:44.808785 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:44 crc kubenswrapper[4678]: I1206 10:36:44.920739 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:45 crc kubenswrapper[4678]: I1206 10:36:45.119270 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:45 crc kubenswrapper[4678]: E1206 10:36:45.545838 4678 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 06 10:36:45 crc kubenswrapper[4678]: I1206 10:36:45.571605 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:45 crc kubenswrapper[4678]: I1206 10:36:45.572863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:45 crc kubenswrapper[4678]: I1206 10:36:45.572902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:45 crc kubenswrapper[4678]: I1206 10:36:45.572916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.574415 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.575988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.576070 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.576098 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.579449 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.605720 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.606014 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.607789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.607840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:46 crc kubenswrapper[4678]: I1206 10:36:46.607855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:47 crc kubenswrapper[4678]: I1206 10:36:47.576347 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:47 crc kubenswrapper[4678]: I1206 10:36:47.577444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:47 crc kubenswrapper[4678]: I1206 10:36:47.577470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:47 crc kubenswrapper[4678]: I1206 10:36:47.577480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:47 crc kubenswrapper[4678]: I1206 10:36:47.920801 4678 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 10:36:47 crc kubenswrapper[4678]: I1206 10:36:47.920924 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:36:48 crc kubenswrapper[4678]: I1206 10:36:48.396897 4678 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 10:36:48 crc kubenswrapper[4678]: I1206 10:36:48.397003 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 10:36:48 crc kubenswrapper[4678]: I1206 10:36:48.418305 4678 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 06 10:36:48 crc kubenswrapper[4678]: I1206 10:36:48.418378 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.560137 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.560837 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.562504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.562580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.562593 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.570982 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.582997 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.583070 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.584182 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.584228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:50 crc kubenswrapper[4678]: I1206 10:36:50.584246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:53 crc kubenswrapper[4678]: E1206 10:36:53.400845 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.406902 4678 trace.go:236] Trace[1861344716]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 10:36:38.535) (total time: 14871ms): Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[1861344716]: ---"Objects listed" error: 14871ms (10:36:53.406) Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[1861344716]: [14.871601833s] [14.871601833s] END Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.406941 4678 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.406913 4678 trace.go:236] Trace[988809406]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 10:36:38.749) (total time: 14657ms): Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[988809406]: ---"Objects listed" error: 14657ms (10:36:53.406) Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[988809406]: [14.657185767s] [14.657185767s] END Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.407010 4678 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.408131 4678 trace.go:236] Trace[351744032]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 10:36:39.183) (total time: 14224ms): Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[351744032]: ---"Objects listed" error: 14224ms (10:36:53.408) Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[351744032]: [14.224993071s] [14.224993071s] END Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.408149 4678 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.409206 4678 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.409548 4678 trace.go:236] Trace[1117883307]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (06-Dec-2025 10:36:38.555) (total time: 14853ms): Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[1117883307]: ---"Objects listed" error: 14853ms (10:36:53.409) Dec 06 10:36:53 crc kubenswrapper[4678]: Trace[1117883307]: [14.853618984s] [14.853618984s] END Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.409579 4678 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 10:36:53 crc kubenswrapper[4678]: E1206 10:36:53.419220 4678 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.436472 4678 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.460126 4678 csr.go:261] certificate signing request csr-nzzjl is approved, waiting to be issued Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.465425 4678 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45206->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.465567 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45206->192.168.126.11:17697: read: connection reset by peer" Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.465432 4678 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45220->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.465718 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45220->192.168.126.11:17697: read: connection reset by peer" Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.465996 4678 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.466065 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.471065 4678 csr.go:257] certificate signing request csr-nzzjl is issued Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.596104 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.598239 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d" exitCode=255 Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.598295 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d"} Dec 06 10:36:53 crc kubenswrapper[4678]: I1206 10:36:53.667466 4678 scope.go:117] "RemoveContainer" containerID="836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.399574 4678 apiserver.go:52] "Watching apiserver" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.402191 4678 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.402600 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.403106 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.403359 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.403460 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.403361 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.403544 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.403581 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.403631 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.403602 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.403751 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.404000 4678 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.407749 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.408759 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.408814 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.408848 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.408934 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.412313 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.412321 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.412600 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.412882 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413585 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413620 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413646 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413667 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413684 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413753 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413775 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413797 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413822 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413844 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413865 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413889 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413905 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413942 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413960 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413979 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.413997 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414016 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414035 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414053 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414070 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414087 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414104 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414121 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414138 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414158 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414177 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414195 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414212 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414232 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414253 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414270 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414288 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414296 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414308 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414383 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414407 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414428 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414446 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414463 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414500 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414516 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414531 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414548 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414567 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414587 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414592 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414603 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414625 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414643 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414659 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414677 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414695 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414712 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414732 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414753 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414772 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414789 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414806 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414827 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414848 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414865 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414884 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414902 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414922 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414941 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414958 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414977 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.414996 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415013 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415034 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415053 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415074 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415091 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415109 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415141 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415156 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415172 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415190 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415207 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415224 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415240 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415255 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415275 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415298 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415321 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415350 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415374 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415397 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415415 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415430 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415446 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415465 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415481 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415521 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415539 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415557 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415574 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415592 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415611 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415658 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415662 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415711 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415736 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415757 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415778 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415796 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415820 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415837 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415855 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415872 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415897 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415917 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415935 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.415986 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416008 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416024 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416041 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416058 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416074 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416092 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416108 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416125 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416141 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416161 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416199 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416308 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416338 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416355 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416372 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416393 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416417 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416442 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416462 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416477 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416526 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416813 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416834 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416852 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416869 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416887 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416903 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416922 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416941 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416963 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416980 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416987 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.416997 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417018 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417038 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417057 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417079 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417099 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417129 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417151 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417171 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417189 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417209 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417226 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417237 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417243 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417333 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417357 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417382 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417445 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417468 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417507 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417527 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417552 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417573 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417602 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417623 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417646 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417673 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417693 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417713 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417733 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417751 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417771 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417790 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417809 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417828 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417849 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417869 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417893 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417912 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417903 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417930 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417955 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417975 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.417995 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418014 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418035 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418053 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418071 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418089 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418110 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418112 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418131 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418155 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418176 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418199 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418220 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418258 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418265 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418290 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418315 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418339 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418365 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418387 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418409 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418435 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418456 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418460 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418474 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418527 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418550 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418570 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418589 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418734 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.418795 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.418970 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419006 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419037 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419065 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.419170 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:54.919140345 +0000 UTC m=+19.762571784 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419203 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419226 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419246 4678 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419262 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.419279 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.421580 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.421968 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.423458 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.424338 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.425107 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.425914 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.426367 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.426693 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.426942 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.427874 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.429373 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.430624 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.431649 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.432871 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.433790 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.433892 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.434083 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.434315 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.434341 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.434591 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.434637 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.435095 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.435256 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.435264 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.435307 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.435738 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.436131 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.438970 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.444821 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.445006 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.445329 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.445366 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.445538 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.445731 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.446301 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.446865 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.446933 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.447130 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.447467 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.447585 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.447641 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.447981 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.448009 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.448248 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.448445 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.448622 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.448894 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.448939 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.449239 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.449350 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.448352 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.449848 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450070 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450205 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450313 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450614 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450733 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450823 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450880 4678 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.450888 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.454580 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451013 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451205 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451265 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451375 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451432 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451666 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451750 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.451970 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.452098 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.452162 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.452213 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.452448 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.452516 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.452760 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.452851 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.453056 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.453342 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.453471 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.453602 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.453727 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.454005 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.454175 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.454236 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.453512 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.454445 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.455039 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:36:54.955016671 +0000 UTC m=+19.798448110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.455139 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.455294 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.455542 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.455604 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.456802 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.456996 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.457405 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.457688 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.457787 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.456864 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-fqd9d"] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.458249 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-qzmlf"] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.458262 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.458454 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.458693 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-z5s25"] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.458836 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-bkhrd"] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.459082 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.459372 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.459281 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.459658 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.459692 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.459870 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460139 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460283 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460343 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460615 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460630 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460669 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460925 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.460974 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.461009 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.461219 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.461387 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.461346 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.461715 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.461711 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.461718 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.462132 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.462191 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.462294 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.462429 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.462719 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.462754 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.462854 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.463083 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.463135 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.463454 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.463513 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.463706 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.464820 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.465078 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.465180 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.465277 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.465368 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.465692 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.467785 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.467934 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.468058 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.468209 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.468358 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.468474 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.468691 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.471440 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.471685 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.471946 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.472250 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-06 10:31:53 +0000 UTC, rotation deadline is 2026-08-24 12:07:05.42642427 +0000 UTC Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.472301 4678 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6265h30m10.954125177s for next certificate rotation Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.472084 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.472511 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.472614 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.472722 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.472971 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.473075 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.473410 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.473618 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.473702 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.473729 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.474109 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.474244 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.474330 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.474478 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.475134 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.475201 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:54.975182822 +0000 UTC m=+19.818614261 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.475357 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.475747 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.475838 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.476278 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.476454 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.476699 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.476797 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.476852 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.476920 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.477069 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.477115 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.477201 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.477612 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.477722 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.478287 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.478462 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.478657 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.479109 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.479823 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.482597 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.482844 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.482912 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.483221 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.483257 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.483266 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.483735 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.483830 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.484670 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.485692 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.486247 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.488915 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.489479 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.490231 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.490297 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.490357 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.491306 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.491477 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.492401 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.492435 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.492463 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.492553 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:54.992531677 +0000 UTC m=+19.835963116 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.493104 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.493429 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.493682 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.493725 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.494103 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.494818 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.440085 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.441361 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.497796 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.499317 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.500752 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.501428 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.502388 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.508758 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.511131 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.511161 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.511176 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.511875 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:55.011234274 +0000 UTC m=+19.854665703 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.515348 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.524863 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525059 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525087 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-cni-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525109 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-socket-dir-parent\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525127 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-conf-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525246 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-os-release\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525350 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-os-release\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525377 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-netns\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525422 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-hostroot\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525750 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-system-cni-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.525978 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-kubelet\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.526019 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-daemon-config\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.526093 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6b30be2-60a6-4733-be36-3f70c6ca6b03-mcd-auth-proxy-config\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.526221 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-k8s-cni-cncf-io\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.526257 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqblr\" (UniqueName: \"kubernetes.io/projected/7c41b983-e061-4cf2-94aa-b0926999502e-kube-api-access-kqblr\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.526325 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxkkj\" (UniqueName: \"kubernetes.io/projected/a6b30be2-60a6-4733-be36-3f70c6ca6b03-kube-api-access-pxkkj\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.526408 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-multus-certs\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.526663 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-cnibin\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527389 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a6b30be2-60a6-4733-be36-3f70c6ca6b03-rootfs\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527414 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527441 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6b30be2-60a6-4733-be36-3f70c6ca6b03-proxy-tls\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527461 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzbth\" (UniqueName: \"kubernetes.io/projected/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-kube-api-access-dzbth\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527511 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-system-cni-dir\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527535 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7c41b983-e061-4cf2-94aa-b0926999502e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527552 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-etc-kubernetes\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527571 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-cnibin\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527589 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-cni-binary-copy\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527606 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-cni-bin\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527622 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-cni-multus\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527701 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dktr6\" (UniqueName: \"kubernetes.io/projected/e219fa13-5c09-41c9-b89f-c00cd49c4d52-kube-api-access-dktr6\") pod \"node-resolver-fqd9d\" (UID: \"e219fa13-5c09-41c9-b89f-c00cd49c4d52\") " pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527902 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7c41b983-e061-4cf2-94aa-b0926999502e-cni-binary-copy\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.528924 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e219fa13-5c09-41c9-b89f-c00cd49c4d52-hosts-file\") pod \"node-resolver-fqd9d\" (UID: \"e219fa13-5c09-41c9-b89f-c00cd49c4d52\") " pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529010 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529870 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529895 4678 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529908 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529919 4678 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529929 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529939 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529949 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529959 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529968 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529979 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529990 4678 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.529999 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530011 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530021 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530031 4678 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530040 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530054 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530064 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530076 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530085 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530095 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530106 4678 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530143 4678 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530153 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.527614 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530164 4678 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530245 4678 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530290 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530303 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530315 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530326 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530358 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530371 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530381 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530391 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530400 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530411 4678 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530143 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530437 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530529 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530543 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530554 4678 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530565 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530576 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530586 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530597 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530608 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530619 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530700 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530712 4678 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530722 4678 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530731 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530741 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530750 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530759 4678 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530768 4678 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530778 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530790 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530799 4678 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530809 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530819 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530828 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530844 4678 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530853 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530862 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530872 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530880 4678 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530889 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530898 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530906 4678 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530916 4678 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530926 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530950 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530958 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530967 4678 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530975 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530984 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.530994 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531005 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531015 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531024 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531035 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531044 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531052 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531063 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531072 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531081 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531090 4678 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531100 4678 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531109 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531118 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531126 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531135 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531144 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531152 4678 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531161 4678 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531171 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531181 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531253 4678 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531267 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531276 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531286 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531296 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531304 4678 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531313 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531323 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531334 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531343 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531353 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531361 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531371 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531379 4678 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531537 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531581 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531596 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531611 4678 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531647 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531665 4678 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531678 4678 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531692 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531723 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531737 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531751 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531763 4678 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531774 4678 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531805 4678 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531820 4678 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531832 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531844 4678 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531856 4678 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531886 4678 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531898 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531911 4678 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531925 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531937 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531949 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531978 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.531991 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532004 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532020 4678 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532031 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532045 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532057 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532069 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532081 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532094 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532138 4678 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532152 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532165 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532177 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532189 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532201 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532213 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532226 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532239 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532252 4678 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532265 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532277 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532290 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532304 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532315 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532346 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532360 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532374 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532387 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532401 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532414 4678 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532433 4678 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532447 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532459 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532472 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532500 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532514 4678 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532528 4678 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532541 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532553 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532566 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532580 4678 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532592 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532604 4678 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532616 4678 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532627 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532638 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532650 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.532662 4678 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.539346 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.546034 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.547987 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.549785 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.562617 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.572400 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.584543 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.596281 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.609383 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.616184 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.619423 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf"} Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.619827 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.620579 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.621176 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2a15f8691436b1e5bda829e610a1ef1ceea058583169a515f11cb72d5a259897"} Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.632582 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633252 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-conf-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633312 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633339 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-cni-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633359 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-socket-dir-parent\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633380 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-os-release\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633380 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-conf-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633415 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-os-release\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633437 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-netns\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633457 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-hostroot\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633466 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-cni-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633477 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-system-cni-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633537 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-kubelet\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633561 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-daemon-config\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633586 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6b30be2-60a6-4733-be36-3f70c6ca6b03-mcd-auth-proxy-config\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633596 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-socket-dir-parent\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633608 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-k8s-cni-cncf-io\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633634 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqblr\" (UniqueName: \"kubernetes.io/projected/7c41b983-e061-4cf2-94aa-b0926999502e-kube-api-access-kqblr\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633660 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxkkj\" (UniqueName: \"kubernetes.io/projected/a6b30be2-60a6-4733-be36-3f70c6ca6b03-kube-api-access-pxkkj\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633682 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-multus-certs\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633704 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-cnibin\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633771 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a6b30be2-60a6-4733-be36-3f70c6ca6b03-rootfs\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633797 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6b30be2-60a6-4733-be36-3f70c6ca6b03-proxy-tls\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-os-release\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633817 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzbth\" (UniqueName: \"kubernetes.io/projected/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-kube-api-access-dzbth\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633841 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-system-cni-dir\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633850 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-os-release\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633863 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7c41b983-e061-4cf2-94aa-b0926999502e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633888 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-etc-kubernetes\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633921 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dktr6\" (UniqueName: \"kubernetes.io/projected/e219fa13-5c09-41c9-b89f-c00cd49c4d52-kube-api-access-dktr6\") pod \"node-resolver-fqd9d\" (UID: \"e219fa13-5c09-41c9-b89f-c00cd49c4d52\") " pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633942 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-cnibin\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633953 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-netns\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633961 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-cni-binary-copy\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634010 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-cni-bin\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634036 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-cni-multus\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634065 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-hostroot\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634074 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7c41b983-e061-4cf2-94aa-b0926999502e-cni-binary-copy\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634155 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e219fa13-5c09-41c9-b89f-c00cd49c4d52-hosts-file\") pod \"node-resolver-fqd9d\" (UID: \"e219fa13-5c09-41c9-b89f-c00cd49c4d52\") " pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634242 4678 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634255 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634325 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-cni-bin\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634389 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-cni-multus\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634664 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-cni-binary-copy\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.634963 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/e219fa13-5c09-41c9-b89f-c00cd49c4d52-hosts-file\") pod \"node-resolver-fqd9d\" (UID: \"e219fa13-5c09-41c9-b89f-c00cd49c4d52\") " pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.635031 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-cnibin\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.635088 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-var-lib-kubelet\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.635226 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-system-cni-dir\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.635275 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-k8s-cni-cncf-io\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.635357 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/a6b30be2-60a6-4733-be36-3f70c6ca6b03-rootfs\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.633561 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-system-cni-dir\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.635685 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-etc-kubernetes\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.635731 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7c41b983-e061-4cf2-94aa-b0926999502e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.636182 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7c41b983-e061-4cf2-94aa-b0926999502e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.636245 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-cnibin\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.636274 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-host-run-multus-certs\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.636896 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6b30be2-60a6-4733-be36-3f70c6ca6b03-mcd-auth-proxy-config\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.637998 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7c41b983-e061-4cf2-94aa-b0926999502e-cni-binary-copy\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.638875 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-multus-daemon-config\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.645737 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a6b30be2-60a6-4733-be36-3f70c6ca6b03-proxy-tls\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.647372 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.652399 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxkkj\" (UniqueName: \"kubernetes.io/projected/a6b30be2-60a6-4733-be36-3f70c6ca6b03-kube-api-access-pxkkj\") pod \"machine-config-daemon-bkhrd\" (UID: \"a6b30be2-60a6-4733-be36-3f70c6ca6b03\") " pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.654855 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzbth\" (UniqueName: \"kubernetes.io/projected/388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c-kube-api-access-dzbth\") pod \"multus-z5s25\" (UID: \"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\") " pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.656599 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqblr\" (UniqueName: \"kubernetes.io/projected/7c41b983-e061-4cf2-94aa-b0926999502e-kube-api-access-kqblr\") pod \"multus-additional-cni-plugins-qzmlf\" (UID: \"7c41b983-e061-4cf2-94aa-b0926999502e\") " pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.663511 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dktr6\" (UniqueName: \"kubernetes.io/projected/e219fa13-5c09-41c9-b89f-c00cd49c4d52-kube-api-access-dktr6\") pod \"node-resolver-fqd9d\" (UID: \"e219fa13-5c09-41c9-b89f-c00cd49c4d52\") " pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.663710 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.688067 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.707415 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.719234 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.719454 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.741313 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.755810 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.771801 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.794811 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.795403 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.814937 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5lfqt"] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.815814 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.823258 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.823525 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.823619 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.823650 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.823823 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.826789 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fqd9d" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.831015 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.831266 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.831429 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.839751 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.841039 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" Dec 06 10:36:54 crc kubenswrapper[4678]: W1206 10:36:54.846068 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6b30be2_60a6_4733_be36_3f70c6ca6b03.slice/crio-22548288da658f00146ef1a1a0ba79563b851460b51758ed2f76ed160c7b88ae WatchSource:0}: Error finding container 22548288da658f00146ef1a1a0ba79563b851460b51758ed2f76ed160c7b88ae: Status 404 returned error can't find the container with id 22548288da658f00146ef1a1a0ba79563b851460b51758ed2f76ed160c7b88ae Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.846684 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-z5s25" Dec 06 10:36:54 crc kubenswrapper[4678]: W1206 10:36:54.853574 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode219fa13_5c09_41c9_b89f_c00cd49c4d52.slice/crio-1b8c60b813bdb4bcbf97e21758ecb07f2e635d048a5029e2c004db952c4e910e WatchSource:0}: Error finding container 1b8c60b813bdb4bcbf97e21758ecb07f2e635d048a5029e2c004db952c4e910e: Status 404 returned error can't find the container with id 1b8c60b813bdb4bcbf97e21758ecb07f2e635d048a5029e2c004db952c4e910e Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.859477 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.884028 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.905458 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.921416 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.934146 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.935877 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.935925 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-ovn-kubernetes\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.935972 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-bin\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.935993 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-systemd-units\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936012 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-ovn\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936029 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-log-socket\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936050 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-config\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936069 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovn-node-metrics-cert\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936108 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-kubelet\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936128 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-etc-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936147 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjjj4\" (UniqueName: \"kubernetes.io/projected/c5cca846-f84b-4924-b1c6-4ec6cea71a65-kube-api-access-qjjj4\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936166 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936183 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-systemd\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936201 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-slash\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936220 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-var-lib-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936254 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936275 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-script-lib\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936296 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-netd\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936317 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-netns\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936334 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-node-log\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936354 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-env-overrides\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.936593 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: E1206 10:36:54.936648 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:55.936628604 +0000 UTC m=+20.780060043 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.936812 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.947856 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.949798 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.949806 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.959638 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.969416 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.980901 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:54 crc kubenswrapper[4678]: I1206 10:36:54.990378 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.005921 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.026069 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037186 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037268 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-config\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037296 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovn-node-metrics-cert\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037316 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037332 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037359 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-kubelet\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037732 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-etc-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037753 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjjj4\" (UniqueName: \"kubernetes.io/projected/c5cca846-f84b-4924-b1c6-4ec6cea71a65-kube-api-access-qjjj4\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037769 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037787 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-systemd\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037804 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037825 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-slash\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037841 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-var-lib-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037855 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-script-lib\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037870 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-netd\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037886 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-netns\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037900 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-node-log\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037913 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-env-overrides\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037932 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037949 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-ovn-kubernetes\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037963 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-bin\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037977 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-systemd-units\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037989 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-ovn\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038004 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-log-socket\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038040 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-log-socket\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.037456 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038099 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:56.03808397 +0000 UTC m=+20.881515409 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038331 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-etc-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.037619 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-kubelet\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038479 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038535 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-systemd-units\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038543 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-node-log\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038562 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038587 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-systemd\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038605 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-ovn-kubernetes\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038626 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-bin\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038663 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038676 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038687 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038719 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:56.03871031 +0000 UTC m=+20.882141749 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038744 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-netd\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038767 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-netns\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.037705 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038785 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038792 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.038812 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:56.038804823 +0000 UTC m=+20.882236262 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038810 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-slash\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038790 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-ovn\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.038843 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-var-lib-openvswitch\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.042587 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-config\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.043078 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:36:56.043009163 +0000 UTC m=+20.886440612 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.043550 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-script-lib\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.044327 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.044380 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovn-node-metrics-cert\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.045388 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-env-overrides\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.060201 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjjj4\" (UniqueName: \"kubernetes.io/projected/c5cca846-f84b-4924-b1c6-4ec6cea71a65-kube-api-access-qjjj4\") pod \"ovnkube-node-5lfqt\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.062274 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.074290 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.080909 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.089957 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.099236 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.114437 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.133175 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.140722 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.153937 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5cca846_f84b_4924_b1c6_4ec6cea71a65.slice/crio-96ba415b0bee77024b2bb13f377a61308cd4a59a064d968e1201ee85e1c014df WatchSource:0}: Error finding container 96ba415b0bee77024b2bb13f377a61308cd4a59a064d968e1201ee85e1c014df: Status 404 returned error can't find the container with id 96ba415b0bee77024b2bb13f377a61308cd4a59a064d968e1201ee85e1c014df Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.155471 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.167540 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.180023 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.191985 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.213691 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.228815 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.246362 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.305043 4678 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305368 4678 reflector.go:484] object-"openshift-ovn-kubernetes"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305426 4678 reflector.go:484] object-"openshift-network-node-identity"/"env-overrides": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"env-overrides": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305439 4678 reflector.go:484] object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305543 4678 reflector.go:484] object-"openshift-network-node-identity"/"network-node-identity-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-network-node-identity"/"network-node-identity-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305563 4678 reflector.go:484] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305579 4678 reflector.go:484] object-"openshift-multus"/"multus-daemon-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"multus-daemon-config": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305609 4678 reflector.go:484] object-"openshift-multus"/"default-cni-sysctl-allowlist": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"default-cni-sysctl-allowlist": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305642 4678 reflector.go:484] object-"openshift-multus"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305649 4678 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305672 4678 reflector.go:484] object-"openshift-network-operator"/"metrics-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-network-operator"/"metrics-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305686 4678 reflector.go:484] object-"openshift-network-operator"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305728 4678 reflector.go:484] object-"openshift-network-operator"/"iptables-alerter-script": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"iptables-alerter-script": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305758 4678 reflector.go:484] object-"openshift-dns"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305785 4678 reflector.go:484] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": watch of *v1.Secret ended with: very short watch: object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305816 4678 reflector.go:484] object-"openshift-machine-config-operator"/"kube-rbac-proxy": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-rbac-proxy": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305855 4678 reflector.go:484] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305889 4678 reflector.go:484] object-"openshift-multus"/"cni-copy-resources": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"cni-copy-resources": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305916 4678 reflector.go:484] object-"openshift-dns"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-dns"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305925 4678 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": watch of *v1.Secret ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305946 4678 reflector.go:484] object-"openshift-machine-config-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-machine-config-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305955 4678 reflector.go:484] object-"openshift-network-operator"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-operator"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305972 4678 reflector.go:484] object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.306028 4678 reflector.go:484] object-"openshift-network-node-identity"/"kube-root-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"kube-root-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.306090 4678 reflector.go:484] object-"openshift-multus"/"default-dockercfg-2q5b6": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"default-dockercfg-2q5b6": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.306119 4678 reflector.go:484] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": watch of *v1.Secret ended with: very short watch: object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305893 4678 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-script-lib": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305461 4678 reflector.go:484] object-"openshift-network-node-identity"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.306316 4678 reflector.go:484] object-"openshift-ovn-kubernetes"/"ovnkube-config": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-ovn-kubernetes"/"ovnkube-config": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305502 4678 reflector.go:484] object-"openshift-multus"/"openshift-service-ca.crt": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-multus"/"openshift-service-ca.crt": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.306354 4678 reflector.go:484] object-"openshift-network-node-identity"/"ovnkube-identity-cm": watch of *v1.ConfigMap ended with: very short watch: object-"openshift-network-node-identity"/"ovnkube-identity-cm": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: W1206 10:36:55.305608 4678 reflector.go:484] object-"openshift-machine-config-operator"/"proxy-tls": watch of *v1.Secret ended with: very short watch: object-"openshift-machine-config-operator"/"proxy-tls": Unexpected watch close - watch lasted less than a second and no items received Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.475603 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.475800 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.482704 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.483349 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.484911 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.485638 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.487076 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.487842 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.488576 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.489813 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.490598 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.491814 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.492511 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.494670 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.495329 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.495711 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.496006 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.497146 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.497843 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.499153 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.500216 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.501010 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.501759 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.502455 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.503136 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.503966 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.505552 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.506074 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.507187 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.507948 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.508960 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.509709 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.510622 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.511142 4678 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.511319 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.513520 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.514064 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.514530 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.516180 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.517322 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.517930 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.519225 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.519821 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.520479 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.521541 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.522172 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.524088 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.524827 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.525315 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.525894 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.526478 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.527462 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.527974 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.528423 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.529024 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.529755 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.530537 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.531115 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.534107 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.551996 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.564789 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.578448 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.593388 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.611599 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.625542 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c41b983-e061-4cf2-94aa-b0926999502e" containerID="196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614" exitCode=0 Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.625600 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerDied","Data":"196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.625640 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerStarted","Data":"9b19050d75f31d303ddd04c1e9b4a4769da01416556a801bd1ad45dca8b09700"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.628505 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fqd9d" event={"ID":"e219fa13-5c09-41c9-b89f-c00cd49c4d52","Type":"ContainerStarted","Data":"36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.628563 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fqd9d" event={"ID":"e219fa13-5c09-41c9-b89f-c00cd49c4d52","Type":"ContainerStarted","Data":"1b8c60b813bdb4bcbf97e21758ecb07f2e635d048a5029e2c004db952c4e910e"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.631468 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.636524 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.636592 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.636606 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"22548288da658f00146ef1a1a0ba79563b851460b51758ed2f76ed160c7b88ae"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.638856 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.638897 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f48fc4bb541f7c59b49eb37fa89ed555219f93aadc3f2aa08206f0a0cc6a7a01"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.649768 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.649823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.652852 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5" exitCode=0 Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.652940 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.653672 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"96ba415b0bee77024b2bb13f377a61308cd4a59a064d968e1201ee85e1c014df"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.654917 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.657143 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerStarted","Data":"34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.657207 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerStarted","Data":"2387b9c16b24588f28d8c307ed100b7ed4d49fe21f22771bde124104646f6559"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.659066 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9fe4a178ea5f6a5465737785aac2f29aca0d11571cd8f26dbf47e7bd4111e097"} Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.673321 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.715084 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.756967 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.795956 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.831286 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.876193 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.929815 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.947899 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.948059 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: E1206 10:36:55.948129 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:57.948107737 +0000 UTC m=+22.791539176 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:55 crc kubenswrapper[4678]: I1206 10:36:55.952328 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.004111 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.033803 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.048440 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.048562 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.048598 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048667 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048688 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:36:58.048648395 +0000 UTC m=+22.892079834 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.048723 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048729 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:58.048719537 +0000 UTC m=+22.892150976 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048763 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048799 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048815 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048838 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048853 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048864 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048882 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:58.048854222 +0000 UTC m=+22.892285661 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.048901 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:36:58.048894123 +0000 UTC m=+22.892325562 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.074037 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.120024 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.156113 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.164552 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.183321 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.202011 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.241447 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.261667 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.291253 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.303380 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.305578 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-dvc6z"] Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.305953 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.344386 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.351473 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjcnv\" (UniqueName: \"kubernetes.io/projected/3daabe49-ee26-4bc3-acef-3d376a17fcd8-kube-api-access-tjcnv\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.351542 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3daabe49-ee26-4bc3-acef-3d376a17fcd8-serviceca\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.351589 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3daabe49-ee26-4bc3-acef-3d376a17fcd8-host\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.362192 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.383759 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.401446 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.422626 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.452695 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3daabe49-ee26-4bc3-acef-3d376a17fcd8-host\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.452798 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjcnv\" (UniqueName: \"kubernetes.io/projected/3daabe49-ee26-4bc3-acef-3d376a17fcd8-kube-api-access-tjcnv\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.452820 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3daabe49-ee26-4bc3-acef-3d376a17fcd8-serviceca\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.452839 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3daabe49-ee26-4bc3-acef-3d376a17fcd8-host\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.454300 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3daabe49-ee26-4bc3-acef-3d376a17fcd8-serviceca\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.461920 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.465891 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.475272 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.475435 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.475828 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.475876 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.481680 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.501515 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.565443 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.583266 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.602694 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.619960 4678 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.622073 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.622103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.622112 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.622222 4678 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.623744 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.634811 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.641804 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.648841 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.662069 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.664900 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e"} Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.664946 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880"} Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.664959 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a"} Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.664971 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30"} Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.664983 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1"} Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.666853 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c41b983-e061-4cf2-94aa-b0926999502e" containerID="168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c" exitCode=0 Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.667547 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerDied","Data":"168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c"} Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.670585 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjcnv\" (UniqueName: \"kubernetes.io/projected/3daabe49-ee26-4bc3-acef-3d376a17fcd8-kube-api-access-tjcnv\") pod \"node-ca-dvc6z\" (UID: \"3daabe49-ee26-4bc3-acef-3d376a17fcd8\") " pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.699345 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.701396 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.723031 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.744160 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.763450 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.782674 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.821576 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.855340 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.882703 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.901169 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.921503 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-dvc6z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.921638 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.942162 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.964941 4678 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.965268 4678 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 06 10:36:56 crc kubenswrapper[4678]: W1206 10:36:56.965644 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3daabe49_ee26_4bc3_acef_3d376a17fcd8.slice/crio-7470ec2c493919fc1d1c934180f70b72e577514762be0cea3db1fd8ecc1cb8cc WatchSource:0}: Error finding container 7470ec2c493919fc1d1c934180f70b72e577514762be0cea3db1fd8ecc1cb8cc: Status 404 returned error can't find the container with id 7470ec2c493919fc1d1c934180f70b72e577514762be0cea3db1fd8ecc1cb8cc Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.967193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.967234 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.967246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.967267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.967285 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:56Z","lastTransitionTime":"2025-12-06T10:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.982738 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 10:36:56 crc kubenswrapper[4678]: E1206 10:36:56.989298 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:56Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.997169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.997209 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.997219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.997235 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:56 crc kubenswrapper[4678]: I1206 10:36:56.997245 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:56Z","lastTransitionTime":"2025-12-06T10:36:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.001633 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.009776 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.012976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.013015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.013026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.013042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.013051 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.024964 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.028206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.028269 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.028281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.028296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.028307 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.031714 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.042163 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.042381 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.045943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.045980 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.045992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.046015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.046029 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.058795 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.058950 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.060697 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.060740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.060751 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.060769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.060782 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.061409 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.081951 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.136713 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.163172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.163231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.163244 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.163264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.163278 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.172384 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.211618 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.249904 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.266786 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.266822 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.266832 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.266849 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.266859 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.290313 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.331712 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.369329 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.369768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.369785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.369810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.369826 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.373077 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.415718 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.451449 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.472164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.472234 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.472246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.472263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.472278 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.475764 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.475977 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.493937 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.532039 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.575748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.575797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.575813 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.575833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.575846 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.583366 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.613642 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.680752 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.680810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.680829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.680857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.680877 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.682518 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.686011 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c41b983-e061-4cf2-94aa-b0926999502e" containerID="c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41" exitCode=0 Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.686090 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerDied","Data":"c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.690946 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dvc6z" event={"ID":"3daabe49-ee26-4bc3-acef-3d376a17fcd8","Type":"ContainerStarted","Data":"094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.690991 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-dvc6z" event={"ID":"3daabe49-ee26-4bc3-acef-3d376a17fcd8","Type":"ContainerStarted","Data":"7470ec2c493919fc1d1c934180f70b72e577514762be0cea3db1fd8ecc1cb8cc"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.706627 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.734111 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.750599 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.771143 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.785097 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.785139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.785148 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.785166 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.785178 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.811211 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.854344 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.888682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.888728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.888740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.888758 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.888769 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.890665 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.932257 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.974451 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:57Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.980174 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.980455 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:57 crc kubenswrapper[4678]: E1206 10:36:57.980636 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:01.980618767 +0000 UTC m=+26.824050206 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.990610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.990652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.990661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.990676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:57 crc kubenswrapper[4678]: I1206 10:36:57.990686 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:57Z","lastTransitionTime":"2025-12-06T10:36:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.017200 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.049796 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.080828 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.080954 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081049 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:37:02.081019752 +0000 UTC m=+26.924451231 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081114 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.081117 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081132 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081146 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.081160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081191 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:02.081181287 +0000 UTC m=+26.924612726 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081294 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081317 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081334 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081381 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:02.081366082 +0000 UTC m=+26.924797561 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081433 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.081480 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:02.081467586 +0000 UTC m=+26.924899055 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.093060 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.093103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.093121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.093139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.093150 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.093266 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.134544 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.176796 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.196457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.196520 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.196533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.196550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.196563 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.214143 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.257197 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.297213 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.299594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.299640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.299653 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.299670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.299682 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.330160 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.373047 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.402461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.402543 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.402589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.402614 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.402632 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.419676 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.451416 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.475317 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.475555 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.475647 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:36:58 crc kubenswrapper[4678]: E1206 10:36:58.475753 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.489713 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.508966 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.509570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.509585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.509603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.509614 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.544761 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.571356 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.610656 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.612596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.612655 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.612677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.612706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.612731 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.650611 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.693782 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.698004 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c41b983-e061-4cf2-94aa-b0926999502e" containerID="2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792" exitCode=0 Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.698091 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerDied","Data":"2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.699818 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.718072 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.718124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.718134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.718157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.718170 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.734324 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.772597 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.814133 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.821628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.821677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.821689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.821707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.821718 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.851213 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.891888 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.924455 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.924526 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.924539 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.924556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.924569 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:58Z","lastTransitionTime":"2025-12-06T10:36:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.934569 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:58 crc kubenswrapper[4678]: I1206 10:36:58.972600 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.016358 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.026376 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.026429 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.026444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.026463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.026478 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.052455 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.093372 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.128689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.128727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.128735 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.128749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.128760 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.136938 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.172621 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.211550 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.230712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.230756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.230773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.230793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.230807 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.261461 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.295161 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.333646 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.333684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.333707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.333723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.333734 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.340845 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.373213 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.415320 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.436851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.436880 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.436888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.436903 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.436912 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.476003 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:36:59 crc kubenswrapper[4678]: E1206 10:36:59.476253 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.539333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.539367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.539377 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.539392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.539403 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.641810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.641858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.641870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.641925 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.641938 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.709808 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.714538 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c41b983-e061-4cf2-94aa-b0926999502e" containerID="1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f" exitCode=0 Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.714823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerDied","Data":"1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.740969 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.744546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.744601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.744617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.744636 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.744650 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.760094 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.779270 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.795536 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.809848 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.822139 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.843110 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.846844 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.846891 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.846906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.846933 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.846947 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.855613 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.868072 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.879167 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.890247 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.903816 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.937349 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.949588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.949634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.949646 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.949663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.949674 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:36:59Z","lastTransitionTime":"2025-12-06T10:36:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:36:59 crc kubenswrapper[4678]: I1206 10:36:59.974608 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:36:59Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.017539 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.052564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.052604 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.052613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.052627 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.052636 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.155352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.155402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.155412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.155428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.155439 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.258204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.258248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.258261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.258282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.258296 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.362100 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.362192 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.362217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.362253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.362277 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.465616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.465685 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.465703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.465729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.465747 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.475907 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.475926 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:00 crc kubenswrapper[4678]: E1206 10:37:00.476125 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:00 crc kubenswrapper[4678]: E1206 10:37:00.476288 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.568299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.569022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.569052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.569121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.569147 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.673052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.673118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.673134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.673162 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.673180 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.726918 4678 generic.go:334] "Generic (PLEG): container finished" podID="7c41b983-e061-4cf2-94aa-b0926999502e" containerID="ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb" exitCode=0 Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.726984 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerDied","Data":"ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.754395 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.781323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.781404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.781428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.781455 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.781476 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.786050 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.806285 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.832603 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.843909 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.857477 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.869985 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.884101 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.885624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.885651 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.885658 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.885672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.885682 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.896401 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.909545 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.927101 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.940844 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.955267 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.965824 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.983566 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:00Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.989320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.989365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.989382 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.989404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:00 crc kubenswrapper[4678]: I1206 10:37:00.989422 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:00Z","lastTransitionTime":"2025-12-06T10:37:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.092850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.092905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.092923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.092947 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.092965 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.195775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.195872 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.195897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.195928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.195951 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.299015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.299068 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.299081 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.299101 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.299115 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.401718 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.401775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.401788 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.401808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.401823 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.475176 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:01 crc kubenswrapper[4678]: E1206 10:37:01.475316 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.503914 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.503956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.503965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.503982 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.503993 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.606797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.606858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.606873 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.606892 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.606906 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.709299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.709353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.709367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.709385 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.709399 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.736009 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.736356 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.742327 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" event={"ID":"7c41b983-e061-4cf2-94aa-b0926999502e","Type":"ContainerStarted","Data":"5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.756152 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.773883 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.785642 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.801326 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.812195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.812241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.812252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.812270 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.812282 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.816405 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.830365 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.845446 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.862941 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.881748 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.894638 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.914805 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.914836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.914845 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.914859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.914869 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:01Z","lastTransitionTime":"2025-12-06T10:37:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.921136 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.934326 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.953278 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.968776 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:01 crc kubenswrapper[4678]: I1206 10:37:01.986379 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:01Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.005473 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.017281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.017323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.017339 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.017359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.017377 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.018982 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.024590 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.024814 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.024906 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:10.02487981 +0000 UTC m=+34.868311269 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.032390 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.057027 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.072288 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.087381 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.108300 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.120580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.120733 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.120850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.120994 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.121107 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.125980 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.126289 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:37:10.126258634 +0000 UTC m=+34.969690103 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.126150 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.126534 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.126716 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.126853 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.126876 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.126939 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:10.126923764 +0000 UTC m=+34.970355213 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.127361 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.127454 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.127512 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:10.127484132 +0000 UTC m=+34.970915581 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.127538 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.127631 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.127661 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.127680 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.127747 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:10.127727259 +0000 UTC m=+34.971158738 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.142814 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.157583 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.170400 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.185260 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.202309 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.224671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.224759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.224823 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.224847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.224866 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.224810 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.244823 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.261604 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.327809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.327861 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.327879 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.327903 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.327920 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.430963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.431331 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.431465 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.431645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.431838 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.475593 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.475826 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.476186 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:02 crc kubenswrapper[4678]: E1206 10:37:02.476534 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.535282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.535703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.535863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.535999 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.536131 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.638782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.638846 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.638862 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.638889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.638925 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.741978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.742023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.742035 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.742070 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.742085 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.745858 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.746575 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.827189 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.845355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.845412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.845427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.845449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.845464 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.847104 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.862649 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.882647 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.901766 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.923560 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.948745 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.948806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.948824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.948856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.948871 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:02Z","lastTransitionTime":"2025-12-06T10:37:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.959931 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:02 crc kubenswrapper[4678]: I1206 10:37:02.976728 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.003939 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:02Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.018540 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:03Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.042183 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:03Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.051450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.051519 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.051532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.051551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.051563 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.072475 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:03Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.090696 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:03Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.103738 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:03Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.127545 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:03Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.139718 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:03Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.153657 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.153704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.153714 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.153731 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.153742 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.257342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.257383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.257392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.257406 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.257415 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.359958 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.360271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.360428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.360523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.360588 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.463783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.464135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.464322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.464516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.464686 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.475175 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:03 crc kubenswrapper[4678]: E1206 10:37:03.475440 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.567517 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.567554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.567563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.567577 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.567587 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.669851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.669920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.669942 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.669971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.669995 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.748864 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.772251 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.772306 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.772321 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.772340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.772354 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.875217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.875270 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.875286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.875310 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.875326 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.978172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.978228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.978241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.978260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:03 crc kubenswrapper[4678]: I1206 10:37:03.978274 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:03Z","lastTransitionTime":"2025-12-06T10:37:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.081470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.081562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.081576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.081597 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.081609 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.184647 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.184725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.184744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.184772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.184795 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.288000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.288053 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.288070 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.288091 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.288107 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.390313 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.390359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.390368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.390383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.390393 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.475099 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.475187 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:04 crc kubenswrapper[4678]: E1206 10:37:04.475282 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:04 crc kubenswrapper[4678]: E1206 10:37:04.475444 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.493065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.493108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.493119 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.493136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.493149 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.596163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.596267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.596286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.596312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.596334 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.699656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.699740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.699765 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.699801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.699827 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.753662 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/0.log" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.757110 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8" exitCode=1 Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.757153 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.758081 4678 scope.go:117] "RemoveContainer" containerID="48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.775180 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.790311 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.802548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.802756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.802902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.803029 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.803142 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.814018 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.853081 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.869552 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.886293 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.905446 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.905508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.905521 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.905538 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.905551 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:04Z","lastTransitionTime":"2025-12-06T10:37:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.907510 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:04Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 10:37:04.174913 5875 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 10:37:04.174967 5875 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 10:37:04.174973 5875 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 10:37:04.174997 5875 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 10:37:04.175003 5875 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 10:37:04.175017 5875 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 10:37:04.175036 5875 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 10:37:04.175045 5875 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 10:37:04.175055 5875 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 10:37:04.175058 5875 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 10:37:04.175074 5875 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 10:37:04.175081 5875 factory.go:656] Stopping watch factory\\\\nI1206 10:37:04.175090 5875 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 10:37:04.175095 5875 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:04.175107 5875 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 10:37:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.926524 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.954663 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.975824 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:04 crc kubenswrapper[4678]: I1206 10:37:04.993168 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:04Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.008365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.008425 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.008446 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.008512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.008530 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.013646 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.030455 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.048823 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.066123 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.111384 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.111800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.111892 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.111975 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.112050 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.215282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.215609 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.215692 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.215766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.215841 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.318570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.318970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.319135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.319271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.319396 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.423046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.423522 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.423679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.423864 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.424010 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.476943 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:05 crc kubenswrapper[4678]: E1206 10:37:05.477162 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.494904 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:04Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 10:37:04.174913 5875 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 10:37:04.174967 5875 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 10:37:04.174973 5875 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 10:37:04.174997 5875 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 10:37:04.175003 5875 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 10:37:04.175017 5875 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 10:37:04.175036 5875 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 10:37:04.175045 5875 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 10:37:04.175055 5875 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 10:37:04.175058 5875 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 10:37:04.175074 5875 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 10:37:04.175081 5875 factory.go:656] Stopping watch factory\\\\nI1206 10:37:04.175090 5875 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 10:37:04.175095 5875 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:04.175107 5875 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 10:37:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.504868 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.520951 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.527037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.527079 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.527090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.527108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.527121 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.536534 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.555423 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.573747 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.597456 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.629665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.629707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.629720 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.629739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.629753 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.652266 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.675271 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.690302 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.700857 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.712992 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.730917 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.732472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.732516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.732526 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.732539 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.732591 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.742936 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.753110 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.762554 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/0.log" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.764672 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.764885 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.776267 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.794614 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:04Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 10:37:04.174913 5875 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 10:37:04.174967 5875 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 10:37:04.174973 5875 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 10:37:04.174997 5875 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 10:37:04.175003 5875 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 10:37:04.175017 5875 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 10:37:04.175036 5875 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 10:37:04.175045 5875 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 10:37:04.175055 5875 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 10:37:04.175058 5875 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 10:37:04.175074 5875 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 10:37:04.175081 5875 factory.go:656] Stopping watch factory\\\\nI1206 10:37:04.175090 5875 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 10:37:04.175095 5875 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:04.175107 5875 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 10:37:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.803829 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.814796 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.826181 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.834963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.835002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.835012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.835025 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.835035 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.844610 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.860997 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.880190 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.900190 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.918878 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.930549 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.939248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.939437 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.939523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.939598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.939670 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:05Z","lastTransitionTime":"2025-12-06T10:37:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.943031 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.960463 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:05 crc kubenswrapper[4678]: I1206 10:37:05.990723 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.005713 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.042767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.042816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.042829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.042847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.042880 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.144951 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.144985 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.145061 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.145105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.145115 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.248151 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.248583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.248673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.248776 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.248877 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.352077 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.352142 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.352165 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.352197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.352223 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.455703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.455785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.455810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.455840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.455864 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.458389 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql"] Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.459072 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.461945 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.462125 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.474921 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.474927 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:06 crc kubenswrapper[4678]: E1206 10:37:06.475106 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:06 crc kubenswrapper[4678]: E1206 10:37:06.475182 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.486207 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.499986 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.516008 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.531609 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.546040 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.558018 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.559010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.559046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.559077 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.559097 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.559109 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.572247 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.576025 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/25d9d216-9317-4d16-a1b6-85cd6d1e2543-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.576081 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/25d9d216-9317-4d16-a1b6-85cd6d1e2543-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.576130 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb9ns\" (UniqueName: \"kubernetes.io/projected/25d9d216-9317-4d16-a1b6-85cd6d1e2543-kube-api-access-rb9ns\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.576178 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/25d9d216-9317-4d16-a1b6-85cd6d1e2543-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.583955 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.595251 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.608762 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.629010 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.644226 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.662368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.662421 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.662439 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.662467 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.662485 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.662536 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.677672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/25d9d216-9317-4d16-a1b6-85cd6d1e2543-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.677726 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/25d9d216-9317-4d16-a1b6-85cd6d1e2543-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.677747 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/25d9d216-9317-4d16-a1b6-85cd6d1e2543-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.677787 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb9ns\" (UniqueName: \"kubernetes.io/projected/25d9d216-9317-4d16-a1b6-85cd6d1e2543-kube-api-access-rb9ns\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.678907 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/25d9d216-9317-4d16-a1b6-85cd6d1e2543-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.679160 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/25d9d216-9317-4d16-a1b6-85cd6d1e2543-env-overrides\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.685988 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/25d9d216-9317-4d16-a1b6-85cd6d1e2543-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.686515 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:04Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 10:37:04.174913 5875 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 10:37:04.174967 5875 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 10:37:04.174973 5875 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 10:37:04.174997 5875 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 10:37:04.175003 5875 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 10:37:04.175017 5875 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 10:37:04.175036 5875 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 10:37:04.175045 5875 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 10:37:04.175055 5875 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 10:37:04.175058 5875 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 10:37:04.175074 5875 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 10:37:04.175081 5875 factory.go:656] Stopping watch factory\\\\nI1206 10:37:04.175090 5875 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 10:37:04.175095 5875 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:04.175107 5875 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 10:37:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.696940 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb9ns\" (UniqueName: \"kubernetes.io/projected/25d9d216-9317-4d16-a1b6-85cd6d1e2543-kube-api-access-rb9ns\") pod \"ovnkube-control-plane-749d76644c-p9cql\" (UID: \"25d9d216-9317-4d16-a1b6-85cd6d1e2543\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.700857 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.714112 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.766139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.766177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.766186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.766201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.766212 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.775637 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.776797 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/1.log" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.780134 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/0.log" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.790609 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d" exitCode=1 Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.790663 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.790720 4678 scope.go:117] "RemoveContainer" containerID="48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.792169 4678 scope.go:117] "RemoveContainer" containerID="bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d" Dec 06 10:37:06 crc kubenswrapper[4678]: E1206 10:37:06.792311 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:06 crc kubenswrapper[4678]: W1206 10:37:06.798102 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25d9d216_9317_4d16_a1b6_85cd6d1e2543.slice/crio-9f8abbd944c42fbc10242a6337ce975a307dd103df8e246136e970f07ef0303e WatchSource:0}: Error finding container 9f8abbd944c42fbc10242a6337ce975a307dd103df8e246136e970f07ef0303e: Status 404 returned error can't find the container with id 9f8abbd944c42fbc10242a6337ce975a307dd103df8e246136e970f07ef0303e Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.812541 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.824978 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.843281 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.855541 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.867385 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.869759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.869800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.869809 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.869834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.869845 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.878633 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.888586 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.902882 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.917403 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.928789 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.947817 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:04Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 10:37:04.174913 5875 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 10:37:04.174967 5875 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 10:37:04.174973 5875 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 10:37:04.174997 5875 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 10:37:04.175003 5875 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 10:37:04.175017 5875 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 10:37:04.175036 5875 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 10:37:04.175045 5875 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 10:37:04.175055 5875 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 10:37:04.175058 5875 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 10:37:04.175074 5875 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 10:37:04.175081 5875 factory.go:656] Stopping watch factory\\\\nI1206 10:37:04.175090 5875 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 10:37:04.175095 5875 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:04.175107 5875 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 10:37:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.958108 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.969697 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.973371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.974062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.974255 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.974342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.974433 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:06Z","lastTransitionTime":"2025-12-06T10:37:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.984794 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:06 crc kubenswrapper[4678]: I1206 10:37:06.998678 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:06Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.011257 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.077423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.077458 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.077470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.077504 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.077517 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.181342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.181408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.181426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.181472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.181554 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.285524 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.285584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.285601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.285630 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.285649 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.363290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.363330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.363345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.363367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.363382 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.385821 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.390949 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.391005 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.391022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.391049 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.391068 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.410588 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.415698 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.415770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.415810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.415833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.415849 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.435707 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.448945 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.448991 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.449003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.449023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.449037 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.462881 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.468906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.469278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.469559 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.469647 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.469718 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.475769 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.475901 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.483498 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.483894 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.485498 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.485605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.485678 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.485760 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.485820 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.588113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.588517 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.588616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.588704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.588785 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.690997 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.691032 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.691044 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.691061 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.691075 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.799598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.800645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.801095 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.801275 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.801426 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.801114 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" event={"ID":"25d9d216-9317-4d16-a1b6-85cd6d1e2543","Type":"ContainerStarted","Data":"5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.802286 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" event={"ID":"25d9d216-9317-4d16-a1b6-85cd6d1e2543","Type":"ContainerStarted","Data":"5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.802343 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" event={"ID":"25d9d216-9317-4d16-a1b6-85cd6d1e2543","Type":"ContainerStarted","Data":"9f8abbd944c42fbc10242a6337ce975a307dd103df8e246136e970f07ef0303e"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.803654 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/1.log" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.808224 4678 scope.go:117] "RemoveContainer" containerID="bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d" Dec 06 10:37:07 crc kubenswrapper[4678]: E1206 10:37:07.808540 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.834845 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.854686 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.871347 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.886666 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.900542 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.904338 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.904368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.904378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.904397 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.904409 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:07Z","lastTransitionTime":"2025-12-06T10:37:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.915881 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.931463 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.951283 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.973941 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48a1d4c7e27768f45491b64bf53d57849adeb9b329cb04095956724c9593f3d8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:04Z\\\",\\\"message\\\":\\\"d (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1206 10:37:04.174913 5875 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1206 10:37:04.174967 5875 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1206 10:37:04.174973 5875 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1206 10:37:04.174997 5875 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1206 10:37:04.175003 5875 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1206 10:37:04.175017 5875 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1206 10:37:04.175036 5875 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1206 10:37:04.175045 5875 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1206 10:37:04.175055 5875 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1206 10:37:04.175058 5875 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1206 10:37:04.175074 5875 handler.go:208] Removed *v1.Node event handler 7\\\\nI1206 10:37:04.175081 5875 factory.go:656] Stopping watch factory\\\\nI1206 10:37:04.175090 5875 handler.go:208] Removed *v1.Node event handler 2\\\\nI1206 10:37:04.175095 5875 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:04.175107 5875 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1206 10:37:0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:07 crc kubenswrapper[4678]: I1206 10:37:07.985158 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.001681 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:07Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.006624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.006648 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.006673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.006686 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.006696 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.022879 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.037144 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.059601 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.073148 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.086553 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.099679 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.109008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.109053 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.109069 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.109090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.109104 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.110404 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.130852 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.143717 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.157787 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.169023 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.180342 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.193543 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.211340 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.212324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.212379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.212398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.212423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.212440 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.228181 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.254110 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.270211 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.288136 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.298423 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.312616 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.315127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.315185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.315199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.315220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.315232 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.332190 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.352597 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.382614 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.399781 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.415790 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.417791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.417860 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.417878 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.417905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.417923 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.431087 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.447637 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.469910 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.475193 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.475370 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:08 crc kubenswrapper[4678]: E1206 10:37:08.475368 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:08 crc kubenswrapper[4678]: E1206 10:37:08.475792 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.491607 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.512333 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.520792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.520866 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.520885 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.520914 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.520932 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.545033 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.562837 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.584660 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.598588 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.613174 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.622902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.622972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.622986 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.623007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.623020 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.625942 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.636854 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.647308 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.673459 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.722509 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-wfvj4"] Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.723590 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:08 crc kubenswrapper[4678]: E1206 10:37:08.723792 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.724962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.725024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.725044 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.725070 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.725088 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.750796 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.763806 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.779147 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.798706 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcbs6\" (UniqueName: \"kubernetes.io/projected/5fdea9d8-255a-4ec7-8095-4489828f4136-kube-api-access-lcbs6\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.798736 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.798790 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.809565 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.810820 4678 scope.go:117] "RemoveContainer" containerID="bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d" Dec 06 10:37:08 crc kubenswrapper[4678]: E1206 10:37:08.811106 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.825817 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.828612 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.828675 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.828689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.828711 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.828724 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.839535 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.854411 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.866639 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.879519 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.892095 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.900006 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcbs6\" (UniqueName: \"kubernetes.io/projected/5fdea9d8-255a-4ec7-8095-4489828f4136-kube-api-access-lcbs6\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.900090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:08 crc kubenswrapper[4678]: E1206 10:37:08.900560 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:08 crc kubenswrapper[4678]: E1206 10:37:08.900634 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:09.400613015 +0000 UTC m=+34.244044454 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.912806 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.922149 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcbs6\" (UniqueName: \"kubernetes.io/projected/5fdea9d8-255a-4ec7-8095-4489828f4136-kube-api-access-lcbs6\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.931045 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.931399 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.931687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.931918 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.932180 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:08Z","lastTransitionTime":"2025-12-06T10:37:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.931117 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.950550 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.979105 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:08 crc kubenswrapper[4678]: I1206 10:37:08.991002 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.009305 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:09Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.035859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.035938 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.035987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.036015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.036033 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.139355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.139438 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.139461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.139528 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.139555 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.251536 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.251591 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.251609 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.251633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.251650 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.354216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.354257 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.354269 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.354286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.354298 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.451287 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:09 crc kubenswrapper[4678]: E1206 10:37:09.451589 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:09 crc kubenswrapper[4678]: E1206 10:37:09.452078 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:10.45203922 +0000 UTC m=+35.295470699 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.457444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.457527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.457543 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.457566 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.457582 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.475961 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:09 crc kubenswrapper[4678]: E1206 10:37:09.476639 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.560188 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.560253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.560274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.560300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.560322 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.663645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.664066 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.664217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.664367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.664558 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.769639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.770042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.770182 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.770312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.770441 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.874000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.874059 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.874078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.874102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.874119 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.981272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.981352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.981374 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.981398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:09 crc kubenswrapper[4678]: I1206 10:37:09.981410 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:09Z","lastTransitionTime":"2025-12-06T10:37:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.057863 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.058193 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.058360 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:26.058316525 +0000 UTC m=+50.901748004 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.085378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.085454 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.085476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.085555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.085579 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.159306 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.159540 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.159605 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.159719 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:37:26.159661538 +0000 UTC m=+51.003093027 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.159846 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.159881 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.159924 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.159950 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.159962 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:26.159934257 +0000 UTC m=+51.003365756 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.159849 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.160025 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.160052 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:26.160017599 +0000 UTC m=+51.003449078 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.160063 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.160095 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.160169 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:26.160146523 +0000 UTC m=+51.003578092 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.189005 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.189090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.189113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.189147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.189169 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.291684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.291728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.291739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.291755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.291787 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.394538 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.394642 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.394670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.394703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.394727 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.464521 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.464730 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.464830 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:12.464801723 +0000 UTC m=+37.308233192 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.476046 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.476050 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.476064 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.476454 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.476523 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:10 crc kubenswrapper[4678]: E1206 10:37:10.476247 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.498362 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.498433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.498456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.498522 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.498554 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.601848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.601918 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.601941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.601973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.601996 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.705458 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.705607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.705628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.705652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.705670 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.808558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.808601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.808610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.808624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.808634 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.912225 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.912306 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.912328 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.912358 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:10 crc kubenswrapper[4678]: I1206 10:37:10.912385 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:10Z","lastTransitionTime":"2025-12-06T10:37:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.015550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.015618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.015640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.015694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.015722 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.118615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.118664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.118673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.118688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.118698 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.221140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.221180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.221189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.221209 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.221220 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.324236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.324315 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.324333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.324365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.324385 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.427188 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.427264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.427288 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.427319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.427342 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.475638 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:11 crc kubenswrapper[4678]: E1206 10:37:11.475785 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.529541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.529589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.529600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.529619 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.529630 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.632884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.632941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.632956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.632980 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.632995 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.735678 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.736088 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.736198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.736297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.736384 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.839677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.839750 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.839768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.839795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.839815 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.943174 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.943237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.943260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.943290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:11 crc kubenswrapper[4678]: I1206 10:37:11.943316 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:11Z","lastTransitionTime":"2025-12-06T10:37:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.046337 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.046402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.046436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.046454 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.046466 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.149222 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.149627 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.149803 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.149982 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.150196 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.254189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.254270 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.254294 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.254354 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.254380 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.357578 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.357628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.357646 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.357662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.357674 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.463289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.463341 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.463355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.463374 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.463386 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.475594 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:12 crc kubenswrapper[4678]: E1206 10:37:12.475863 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.475683 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:12 crc kubenswrapper[4678]: E1206 10:37:12.476144 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.475683 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:12 crc kubenswrapper[4678]: E1206 10:37:12.476431 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.487684 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:12 crc kubenswrapper[4678]: E1206 10:37:12.487865 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:12 crc kubenswrapper[4678]: E1206 10:37:12.487920 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:16.487901803 +0000 UTC m=+41.331333252 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.567586 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.567686 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.567708 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.567737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.567759 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.671589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.671671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.671691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.671716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.671736 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.774872 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.774949 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.774973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.775002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.775026 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.879157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.879653 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.879830 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.880002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.880150 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.984664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.984749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.984772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.984805 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:12 crc kubenswrapper[4678]: I1206 10:37:12.984828 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:12Z","lastTransitionTime":"2025-12-06T10:37:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.088459 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.088568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.088591 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.088619 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.088642 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.193030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.193106 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.193126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.193151 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.193169 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.297008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.297078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.297103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.297137 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.297161 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.400263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.400365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.400388 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.400414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.400432 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.475226 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:13 crc kubenswrapper[4678]: E1206 10:37:13.475454 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.503329 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.503411 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.503438 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.503481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.503550 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.606615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.606686 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.606711 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.606738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.606761 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.710544 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.710588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.710597 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.710613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.710624 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.814662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.814727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.814744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.814770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.814791 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.918198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.918256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.918280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.918308 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:13 crc kubenswrapper[4678]: I1206 10:37:13.918328 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:13Z","lastTransitionTime":"2025-12-06T10:37:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.021576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.021658 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.021672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.021693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.021707 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.124118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.124198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.124221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.124252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.124278 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.227120 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.227187 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.227211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.227253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.227275 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.330531 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.330583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.330601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.330626 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.330648 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.433879 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.433945 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.433962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.433989 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.434008 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.475461 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.475569 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.475632 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:14 crc kubenswrapper[4678]: E1206 10:37:14.475723 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:14 crc kubenswrapper[4678]: E1206 10:37:14.475867 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:14 crc kubenswrapper[4678]: E1206 10:37:14.476073 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.537540 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.537934 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.538080 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.538211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.538347 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.642163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.642221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.642242 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.642272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.642296 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.746042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.746104 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.746121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.746145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.746165 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.855867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.856357 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.856603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.856820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.857030 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.961093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.961537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.961715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.961910 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:14 crc kubenswrapper[4678]: I1206 10:37:14.962068 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:14Z","lastTransitionTime":"2025-12-06T10:37:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.065483 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.065708 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.065730 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.065757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.065778 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.206901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.206960 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.206980 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.207009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.207030 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.310387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.310453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.310470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.310532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.310559 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.413755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.413832 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.413851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.413875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.413892 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.475802 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:15 crc kubenswrapper[4678]: E1206 10:37:15.476040 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.499959 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.517155 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.517221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.517245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.517279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.517303 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.525914 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.546033 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.562715 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.579556 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.596330 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.608449 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.620050 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.620112 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.620127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.620147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.620160 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.621869 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.634999 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.658781 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.675750 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.691681 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.720925 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.722793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.722868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.722881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.722898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.722910 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.734332 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.748797 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.763537 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.779703 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.826220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.826276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.826295 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.826326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.826346 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.928962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.929001 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.929012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.929030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:15 crc kubenswrapper[4678]: I1206 10:37:15.929040 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:15Z","lastTransitionTime":"2025-12-06T10:37:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.032754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.032829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.032852 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.032881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.032904 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.135769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.135854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.135878 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.135908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.135929 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.239075 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.239194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.239227 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.239267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.239292 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.342031 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.342077 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.342089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.342113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.342128 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.445318 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.445408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.445428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.445458 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.445477 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.475690 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.475690 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:16 crc kubenswrapper[4678]: E1206 10:37:16.475937 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:16 crc kubenswrapper[4678]: E1206 10:37:16.476053 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.476192 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:16 crc kubenswrapper[4678]: E1206 10:37:16.476442 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.534180 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:16 crc kubenswrapper[4678]: E1206 10:37:16.534407 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:16 crc kubenswrapper[4678]: E1206 10:37:16.534601 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:24.534565388 +0000 UTC m=+49.377996857 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.548714 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.548799 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.548823 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.548856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.548881 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.652226 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.652289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.652301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.652319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.652331 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.756078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.756158 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.756193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.756228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.756249 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.859745 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.859826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.859848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.859876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.859896 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.963109 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.963170 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.963187 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.963207 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:16 crc kubenswrapper[4678]: I1206 10:37:16.963222 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:16Z","lastTransitionTime":"2025-12-06T10:37:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.066288 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.066348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.066365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.066387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.066404 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.169328 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.169413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.169432 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.169462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.169480 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.272695 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.272865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.272883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.272915 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.272973 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.375905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.375979 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.375996 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.376026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.376042 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.475912 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:17 crc kubenswrapper[4678]: E1206 10:37:17.476116 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.478435 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.478508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.478524 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.478542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.478555 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.581061 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.581098 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.581107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.581125 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.581139 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.685244 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.685306 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.685323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.685350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.685368 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.711243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.711362 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.711424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.711452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.711474 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: E1206 10:37:17.730003 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:17Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.736534 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.736732 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.736829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.736934 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.737043 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: E1206 10:37:17.762357 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:17Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.768823 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.768891 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.768913 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.769002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.769024 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: E1206 10:37:17.788780 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:17Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.793868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.793928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.793958 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.793992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.794017 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: E1206 10:37:17.817248 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:17Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.823135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.823226 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.823254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.823285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.823304 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: E1206 10:37:17.845215 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:17Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:17 crc kubenswrapper[4678]: E1206 10:37:17.845731 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.847703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.847757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.847781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.847815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.847840 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.950695 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.950777 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.950805 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.950838 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:17 crc kubenswrapper[4678]: I1206 10:37:17.950864 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:17Z","lastTransitionTime":"2025-12-06T10:37:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.054779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.054858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.054880 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.054908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.054928 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.158346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.158415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.158433 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.158461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.158484 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.260901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.260936 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.260947 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.260960 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.260969 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.364010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.364098 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.364116 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.364146 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.364166 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.467137 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.467243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.467262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.467286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.467304 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.474989 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:18 crc kubenswrapper[4678]: E1206 10:37:18.475183 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.475058 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:18 crc kubenswrapper[4678]: E1206 10:37:18.475382 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.475000 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:18 crc kubenswrapper[4678]: E1206 10:37:18.475606 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.570680 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.570724 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.570733 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.570748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.570758 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.674399 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.674520 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.674548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.674584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.674642 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.777292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.777341 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.777351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.777369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.777380 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.880378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.880449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.880472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.880533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.880563 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.984046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.984164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.984186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.984216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:18 crc kubenswrapper[4678]: I1206 10:37:18.984241 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:18Z","lastTransitionTime":"2025-12-06T10:37:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.087537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.087614 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.087644 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.087676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.087698 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.191325 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.191459 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.191529 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.191570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.191596 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.297067 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.298299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.298562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.298755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.299201 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.403754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.404165 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.404261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.404354 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.404441 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.475942 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:19 crc kubenswrapper[4678]: E1206 10:37:19.476159 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.476306 4678 scope.go:117] "RemoveContainer" containerID="bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.507692 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.507814 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.507845 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.507876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.507896 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.611300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.611671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.611756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.611841 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.611923 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.716726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.717259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.717289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.717325 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.717351 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.820965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.821014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.821023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.821043 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.821053 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.854306 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/1.log" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.858544 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.859049 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.883139 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.900339 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.919053 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.924078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.924133 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.924156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.924185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.924206 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:19Z","lastTransitionTime":"2025-12-06T10:37:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.953476 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.976721 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:19 crc kubenswrapper[4678]: I1206 10:37:19.998354 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.018550 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.027729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.027787 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.027800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.027825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.028030 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.037316 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.053241 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.066032 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.085066 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.097554 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.118138 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.130497 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.130540 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.130550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.130567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.130579 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.131405 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.144423 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.157392 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.170242 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.233143 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.233182 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.233193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.233208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.233220 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.335255 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.335328 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.335348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.335376 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.335396 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.438048 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.438096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.438114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.438132 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.438143 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.475906 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:20 crc kubenswrapper[4678]: E1206 10:37:20.476075 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.475911 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:20 crc kubenswrapper[4678]: E1206 10:37:20.476171 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.475906 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:20 crc kubenswrapper[4678]: E1206 10:37:20.476338 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.541003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.541055 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.541065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.541086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.541097 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.644774 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.644848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.644878 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.644909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.644934 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.748914 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.748976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.748987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.749003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.749014 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.852068 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.852116 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.852132 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.852156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.852173 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.864428 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/2.log" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.865108 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/1.log" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.868122 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f" exitCode=1 Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.868175 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.868231 4678 scope.go:117] "RemoveContainer" containerID="bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.869415 4678 scope.go:117] "RemoveContainer" containerID="dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f" Dec 06 10:37:20 crc kubenswrapper[4678]: E1206 10:37:20.869733 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.890466 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.914931 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.931887 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.950089 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.955087 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.955140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.955152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.955586 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.955656 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:20Z","lastTransitionTime":"2025-12-06T10:37:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.962650 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.975285 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:20 crc kubenswrapper[4678]: I1206 10:37:20.989553 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:20Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.018569 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.035368 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.053910 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.058865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.058922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.058935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.058956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.058970 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.075907 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb8cebb06d546782e3ce66c44bfb44aa152fc54272c84c1864d8fafc5e7d451d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:05Z\\\",\\\"message\\\":\\\"mns:[] Mutations:[{Column:policies Mutator:insert Value:{GoSet:[{GoUUID:a5a72d02-1a0f-4f7f-a8c5-6923a1c4274a}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1206 10:37:05.687935 5997 address_set.go:302] New(0d39bc5c-d5b9-432c-81be-2275bce5d7aa/default-network-controller:EgressIP:node-ips:v4:default/a712973235162149816) with []\\\\nI1206 10:37:05.687956 5997 address_set.go:302] New(aa6fc2dc-fab0-4812-b9da-809058e4dcf7/default-network-controller:EgressIP:egressip-served-pods:v4:default/a8519615025667110816) with []\\\\nI1206 10:37:05.687972 5997 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1206 10:37:05.688015 5997 factory.go:1336] Added *v1.Node event handler 7\\\\nI1206 10:37:05.688087 5997 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1206 10:37:05.688406 5997 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1206 10:37:05.688548 5997 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1206 10:37:05.688641 5997 ovnkube.go:599] Stopped ovnkube\\\\nI1206 10:37:05.688690 5997 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1206 10:37:05.688762 5997 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.089654 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.104803 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.128729 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.148596 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.162973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.163011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.163026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.163047 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.163062 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.169723 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.191079 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.266124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.266197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.266218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.266247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.266270 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.369741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.369829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.369846 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.369869 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.369885 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.473824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.473922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.473950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.473987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.474016 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.475131 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:21 crc kubenswrapper[4678]: E1206 10:37:21.475304 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.577994 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.578065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.578090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.578125 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.578148 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.681165 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.681224 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.681238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.681258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.681274 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.784605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.784659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.784671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.784688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.784700 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.875829 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/2.log" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.880579 4678 scope.go:117] "RemoveContainer" containerID="dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f" Dec 06 10:37:21 crc kubenswrapper[4678]: E1206 10:37:21.880815 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.887404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.887478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.887516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.887541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.887561 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.896212 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.911788 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.923807 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.934412 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.954404 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.968627 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.987880 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:21Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.991098 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.991163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.991180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.991206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:21 crc kubenswrapper[4678]: I1206 10:37:21.991226 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:21Z","lastTransitionTime":"2025-12-06T10:37:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.017727 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.033742 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.056032 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.076881 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.094140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.094393 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.094530 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.094631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.094707 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.097729 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.114251 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.130171 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.132424 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.139737 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.150066 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.163918 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.176918 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.193084 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.198611 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.198677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.198694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.198740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.198754 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.213307 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.233567 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.253676 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.279021 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.294884 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.301608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.301854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.301999 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.302176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.302311 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.315898 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.341580 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.368369 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.395716 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.405168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.405238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.405264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.405293 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.405318 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.413399 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.431290 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.445424 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.468008 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.475240 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.475327 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:22 crc kubenswrapper[4678]: E1206 10:37:22.475448 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.475267 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:22 crc kubenswrapper[4678]: E1206 10:37:22.475658 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:22 crc kubenswrapper[4678]: E1206 10:37:22.475791 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.485906 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.503115 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.507956 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.508042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.508069 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.508102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.508128 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.522391 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.537916 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:22Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.610391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.610444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.610461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.610514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.610532 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.713382 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.713481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.713532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.713559 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.713578 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.817554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.817621 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.817640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.817666 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.817685 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.920950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.921015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.921033 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.921058 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:22 crc kubenswrapper[4678]: I1206 10:37:22.921076 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:22Z","lastTransitionTime":"2025-12-06T10:37:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.024221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.024298 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.024317 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.024346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.024367 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.127950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.128021 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.128039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.128065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.128086 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.232319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.232377 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.232396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.232422 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.232441 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.335088 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.335512 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.335629 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.335725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.335818 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.438886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.439605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.439642 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.439697 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.439717 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.475361 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:23 crc kubenswrapper[4678]: E1206 10:37:23.475638 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.543113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.543164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.543180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.543206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.543224 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.646706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.646769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.646782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.646804 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.646820 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.749871 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.749902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.749910 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.749926 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.749936 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.853025 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.853074 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.853086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.853105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.853117 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.957024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.957115 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.957138 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.957171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:23 crc kubenswrapper[4678]: I1206 10:37:23.957193 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:23Z","lastTransitionTime":"2025-12-06T10:37:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.061133 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.061285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.061319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.061351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.061372 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.164402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.164476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.164528 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.164556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.164575 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.268324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.268386 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.268462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.268532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.268558 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.372208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.372258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.372270 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.372291 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.372305 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.474890 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.474941 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.474980 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:24 crc kubenswrapper[4678]: E1206 10:37:24.475043 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:24 crc kubenswrapper[4678]: E1206 10:37:24.475354 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:24 crc kubenswrapper[4678]: E1206 10:37:24.475423 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.475478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.475564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.475589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.475617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.475641 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.578922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.578993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.579018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.579047 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.579071 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.630332 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:24 crc kubenswrapper[4678]: E1206 10:37:24.630636 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:24 crc kubenswrapper[4678]: E1206 10:37:24.630769 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:40.630739515 +0000 UTC m=+65.474170954 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.681338 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.681368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.681376 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.681390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.681398 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.783644 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.783695 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.783706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.783724 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.783737 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.886481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.886523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.886532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.886548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.886559 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.989083 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.989175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.989193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.989218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:24 crc kubenswrapper[4678]: I1206 10:37:24.989236 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:24Z","lastTransitionTime":"2025-12-06T10:37:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.092398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.092471 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.092529 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.092561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.092582 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.195906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.195971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.195998 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.196027 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.196050 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.299979 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.300042 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.300063 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.300091 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.300116 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.404065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.404129 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.404142 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.404161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.404174 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.475296 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:25 crc kubenswrapper[4678]: E1206 10:37:25.475580 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.492451 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.507171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.507219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.507232 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.507248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.507258 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.509304 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.531400 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.554126 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.573551 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.589278 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.604730 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.610391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.610448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.610463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.610525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.610540 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.622184 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.644230 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.659180 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.676210 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.687412 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.698888 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.713918 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.714014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.714164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.714199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.714215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.714224 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.728508 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.743345 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.760464 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.783514 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.816856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.816941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.816952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.816970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.817035 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.920979 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.921040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.921057 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.921082 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:25 crc kubenswrapper[4678]: I1206 10:37:25.921161 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:25Z","lastTransitionTime":"2025-12-06T10:37:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.024114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.024169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.024184 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.024201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.024214 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.127673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.127706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.127715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.127728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.127738 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.150827 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.150955 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.151002 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:58.150987368 +0000 UTC m=+82.994418797 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.230618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.230667 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.230678 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.230694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.230704 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.251270 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.251385 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.251426 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251456 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:37:58.251430363 +0000 UTC m=+83.094861802 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.251482 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251572 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251617 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:58.251608319 +0000 UTC m=+83.095039758 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251741 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251736 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251822 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251775 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251847 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251861 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251925 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:58.251907408 +0000 UTC m=+83.095338887 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.251955 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:37:58.251942069 +0000 UTC m=+83.095373538 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.334022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.334514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.334527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.334545 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.334556 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.437529 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.437612 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.437631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.437659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.437678 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.475198 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.475274 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.475313 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.475379 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.475512 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:26 crc kubenswrapper[4678]: E1206 10:37:26.475556 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.541218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.541302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.541326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.541363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.541385 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.643953 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.643992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.644000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.644014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.644024 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.747056 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.747139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.747161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.747191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.747219 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.850052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.850131 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.850153 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.850186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.850208 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.953477 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.953546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.953556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.953574 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:26 crc kubenswrapper[4678]: I1206 10:37:26.953588 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:26Z","lastTransitionTime":"2025-12-06T10:37:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.056610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.056672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.056684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.056701 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.056717 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.160409 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.160480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.160531 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.160564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.160583 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.263825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.263902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.263922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.263950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.263969 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.367404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.367468 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.367521 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.367550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.367569 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.470198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.470326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.470354 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.470379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.470396 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.475934 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:27 crc kubenswrapper[4678]: E1206 10:37:27.476136 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.573390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.573457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.573482 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.573551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.573574 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.677330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.677398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.677415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.677439 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.677460 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.780281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.780356 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.780373 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.780400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.780419 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.883065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.883105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.883115 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.883131 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.883141 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.981525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.981596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.981623 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.981667 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:27 crc kubenswrapper[4678]: I1206 10:37:27.981694 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:27Z","lastTransitionTime":"2025-12-06T10:37:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.001407 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:27Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.009816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.009957 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.010004 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.010038 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.010081 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.026037 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:28Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.029812 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.029855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.029869 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.029889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.029904 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.044973 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:28Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.051102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.051185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.051208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.051241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.051267 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.067168 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:28Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.071383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.071438 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.071451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.071468 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.071480 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.085893 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:28Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.086046 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.088201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.088261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.088274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.088296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.088309 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.191102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.191168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.191186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.191218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.191244 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.295058 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.295136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.295154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.295178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.295196 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.399110 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.399184 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.399206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.399237 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.399261 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.475061 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.475149 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.475094 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.475323 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.475534 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:28 crc kubenswrapper[4678]: E1206 10:37:28.475620 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.504682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.504738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.504757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.504782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.504801 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.607906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.607973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.607991 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.608017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.608050 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.710753 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.710817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.710834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.710860 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.710878 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.813447 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.813945 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.814113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.814323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.814515 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.918159 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.918195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.918205 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.918219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:28 crc kubenswrapper[4678]: I1206 10:37:28.918232 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:28Z","lastTransitionTime":"2025-12-06T10:37:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.021201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.021240 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.021250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.021267 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.021276 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.124382 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.124455 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.124481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.124542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.124563 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.227402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.227461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.227470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.227503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.227515 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.331161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.331281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.331304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.331366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.331385 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.435450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.435555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.435600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.435631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.435653 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.475901 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:29 crc kubenswrapper[4678]: E1206 10:37:29.476082 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.539560 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.540003 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.540147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.540364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.540701 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.644657 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.645160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.645351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.645530 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.645692 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.748693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.748755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.748776 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.748805 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.748826 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.852445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.852532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.852550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.852575 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.852593 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.956039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.956121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.956143 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.956176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:29 crc kubenswrapper[4678]: I1206 10:37:29.956200 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:29Z","lastTransitionTime":"2025-12-06T10:37:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.059333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.059379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.059394 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.059415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.059431 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.163107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.163164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.163176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.163195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.163206 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.267800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.267856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.267875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.267906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.267925 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.370630 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.370683 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.370694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.370712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.370724 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.473798 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.473858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.473870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.473891 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.473907 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.475198 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:30 crc kubenswrapper[4678]: E1206 10:37:30.475329 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.475678 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.475787 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:30 crc kubenswrapper[4678]: E1206 10:37:30.475956 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:30 crc kubenswrapper[4678]: E1206 10:37:30.476162 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.577224 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.577301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.577323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.577353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.577372 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.682445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.682558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.682583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.682652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.682678 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.786268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.786365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.786383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.786409 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.786453 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.889287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.889349 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.889358 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.889399 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.889410 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.993087 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.993156 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.993174 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.993202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:30 crc kubenswrapper[4678]: I1206 10:37:30.993221 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:30Z","lastTransitionTime":"2025-12-06T10:37:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.095889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.095926 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.095934 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.095963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.095972 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.198896 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.198964 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.198983 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.199012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.199031 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.302376 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.302427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.302440 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.302460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.302473 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.405997 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.406077 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.406102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.406134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.406158 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.475249 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:31 crc kubenswrapper[4678]: E1206 10:37:31.475544 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.509588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.509693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.509714 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.509746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.509767 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.612598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.612682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.612702 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.612732 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.612752 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.717185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.717461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.717654 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.717793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.717928 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.821360 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.821423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.821440 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.821467 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.821516 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.924019 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.924555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.924759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.925010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:31 crc kubenswrapper[4678]: I1206 10:37:31.925195 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:31Z","lastTransitionTime":"2025-12-06T10:37:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.028243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.028727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.028875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.029039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.029169 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.131780 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.131849 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.131867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.131892 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.131909 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.236232 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.236309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.236323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.236367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.236382 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.340571 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.340638 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.340657 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.340689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.340709 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.444381 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.444453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.444468 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.444525 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.444543 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.475618 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.475689 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:32 crc kubenswrapper[4678]: E1206 10:37:32.475746 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:32 crc kubenswrapper[4678]: E1206 10:37:32.475905 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.475948 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:32 crc kubenswrapper[4678]: E1206 10:37:32.476020 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.547615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.547938 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.548112 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.548217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.548314 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.652037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.652101 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.652112 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.652131 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.652145 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.755899 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.755955 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.755969 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.755989 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.756002 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.860140 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.860217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.860241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.860269 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.860290 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.963511 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.963554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.963566 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.963585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:32 crc kubenswrapper[4678]: I1206 10:37:32.963598 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:32Z","lastTransitionTime":"2025-12-06T10:37:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.066088 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.066127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.066138 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.066157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.066169 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.170851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.170927 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.170951 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.170978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.170999 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.274827 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.274896 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.274916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.274946 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.274970 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.378537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.378606 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.378625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.378656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.378676 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.475778 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:33 crc kubenswrapper[4678]: E1206 10:37:33.476351 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.480978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.481013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.481026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.481040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.481053 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.585014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.585479 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.585713 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.586097 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.586410 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.689654 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.689738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.689756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.689775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.689786 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.792965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.793036 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.793054 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.793082 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.793101 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.896007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.896053 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.896063 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.896080 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.896090 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.998909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.998948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.998957 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.998971 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:33 crc kubenswrapper[4678]: I1206 10:37:33.998984 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:33Z","lastTransitionTime":"2025-12-06T10:37:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.102338 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.102405 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.102427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.102454 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.102473 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.204978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.205024 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.205035 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.205052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.205064 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.307566 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.307608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.307617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.307631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.307639 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.410127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.410183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.410199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.410220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.410234 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.474922 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.475015 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:34 crc kubenswrapper[4678]: E1206 10:37:34.475141 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.475179 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:34 crc kubenswrapper[4678]: E1206 10:37:34.475258 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:34 crc kubenswrapper[4678]: E1206 10:37:34.475371 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.512924 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.513674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.513740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.513771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.513788 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.617558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.617623 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.617636 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.617651 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.617661 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.721083 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.721154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.721172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.721202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.721221 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.825561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.825631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.825641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.825661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.825672 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.929613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.929691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.929716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.929748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:34 crc kubenswrapper[4678]: I1206 10:37:34.929814 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:34Z","lastTransitionTime":"2025-12-06T10:37:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.034036 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.034136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.034163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.034202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.034227 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.137729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.138177 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.138381 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.138610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.138778 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.243377 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.243461 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.243482 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.243541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.243563 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.347128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.347183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.347288 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.347315 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.347333 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.450684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.450736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.450767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.450783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.450794 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.476849 4678 scope.go:117] "RemoveContainer" containerID="dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.477072 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:35 crc kubenswrapper[4678]: E1206 10:37:35.477203 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:35 crc kubenswrapper[4678]: E1206 10:37:35.477265 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.513922 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.531813 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.553411 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.553458 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.553527 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.553551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.553603 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.553812 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.574682 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.590137 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.611857 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.631060 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.645147 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.657988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.658386 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.658469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.658568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.658633 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.659565 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.681392 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.693174 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.707305 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.721968 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.735228 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.747902 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.759619 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.761720 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.761767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.761780 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.761802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.761815 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.772510 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.784659 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:35Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.865545 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.865603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.865622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.865649 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.865671 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.968878 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.968935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.968952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.968976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:35 crc kubenswrapper[4678]: I1206 10:37:35.968995 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:35Z","lastTransitionTime":"2025-12-06T10:37:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.072598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.072664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.072681 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.072707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.072725 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.176535 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.176612 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.176633 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.176661 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.176683 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.280294 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.280359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.280379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.280408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.280429 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.383572 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.383626 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.383640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.383689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.383704 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.475209 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.475335 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:36 crc kubenswrapper[4678]: E1206 10:37:36.475425 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.475227 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:36 crc kubenswrapper[4678]: E1206 10:37:36.475603 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:36 crc kubenswrapper[4678]: E1206 10:37:36.475762 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.486462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.486865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.487064 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.487275 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.487462 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.590274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.590342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.590364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.590390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.590409 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.692918 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.693314 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.693415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.693502 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.693573 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.796597 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.796643 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.796656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.796677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.796692 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.900203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.900248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.900261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.900279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:36 crc kubenswrapper[4678]: I1206 10:37:36.900292 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:36Z","lastTransitionTime":"2025-12-06T10:37:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.003373 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.003416 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.003427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.003445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.003459 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.115249 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.115322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.115333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.115352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.115365 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.220030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.220100 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.220118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.220144 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.220163 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.323406 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.323480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.323515 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.323538 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.323555 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.426893 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.427264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.427272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.427288 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.427298 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.476014 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:37 crc kubenswrapper[4678]: E1206 10:37:37.476179 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.530000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.530032 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.530041 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.530056 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.530067 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.635402 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.635452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.635462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.635476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.635521 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.737972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.738034 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.738058 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.738090 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.738112 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.841451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.841533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.841552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.841575 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.841595 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.944673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.944746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.944831 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.944931 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:37 crc kubenswrapper[4678]: I1206 10:37:37.945028 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:37Z","lastTransitionTime":"2025-12-06T10:37:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.047738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.047797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.047807 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.047828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.047842 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.151045 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.151101 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.151122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.151150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.151171 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.254963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.255069 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.255088 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.255119 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.255137 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.358757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.358840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.358858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.358886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.358911 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.433926 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.433986 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.434004 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.434030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.434049 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.452235 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:38Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.456561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.456599 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.456611 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.456630 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.456644 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.472699 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:38Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.475753 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.475758 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.475865 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.476027 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.476266 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.476422 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.479616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.479687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.479700 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.479744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.479759 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.492369 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:38Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.496801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.496884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.496897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.496920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.496932 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.515448 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:38Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.520150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.520290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.520303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.520346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.520362 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.538455 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:38Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:38 crc kubenswrapper[4678]: E1206 10:37:38.538696 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.540968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.541004 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.541017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.541062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.541085 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.644565 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.644636 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.644650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.644670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.645050 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.748716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.748796 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.748825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.748868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.748892 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.852175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.853334 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.853352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.853373 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.853401 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.957149 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.957212 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.957232 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.957255 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:38 crc kubenswrapper[4678]: I1206 10:37:38.957273 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:38Z","lastTransitionTime":"2025-12-06T10:37:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.059855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.059930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.059950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.059974 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.059993 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.163764 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.163840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.163852 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.163870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.163882 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.267740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.267816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.267838 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.267912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.267940 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.370799 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.370833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.370842 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.370857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.370866 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.473178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.473227 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.473241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.473258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.473269 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.475467 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:39 crc kubenswrapper[4678]: E1206 10:37:39.475599 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.576398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.576445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.576458 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.576476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.576510 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.678884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.678933 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.678946 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.678964 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.678976 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.781588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.782026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.782141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.782245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.782325 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.885716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.885762 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.885773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.885791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.885803 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.988404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.988453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.988463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.988480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:39 crc kubenswrapper[4678]: I1206 10:37:39.988508 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:39Z","lastTransitionTime":"2025-12-06T10:37:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.091122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.091153 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.091161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.091175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.091185 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.194000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.194046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.194058 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.194075 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.194085 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.296601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.296655 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.296665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.296683 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.296697 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.399631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.399692 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.399707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.399727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.399745 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.475616 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.475688 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.475652 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:40 crc kubenswrapper[4678]: E1206 10:37:40.475839 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:40 crc kubenswrapper[4678]: E1206 10:37:40.475956 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:40 crc kubenswrapper[4678]: E1206 10:37:40.476090 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.502096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.502143 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.502152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.502169 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.502180 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.605778 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.605850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.605871 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.605898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.605919 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.709193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.709249 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.709261 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.709286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.709300 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.717752 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:40 crc kubenswrapper[4678]: E1206 10:37:40.717887 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:40 crc kubenswrapper[4678]: E1206 10:37:40.717957 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:38:12.717934483 +0000 UTC m=+97.561365922 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.811768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.811810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.811819 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.811834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.811843 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.914403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.914479 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.914526 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.914553 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:40 crc kubenswrapper[4678]: I1206 10:37:40.914570 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:40Z","lastTransitionTime":"2025-12-06T10:37:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.016671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.016722 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.016737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.016756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.016768 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.119876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.119923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.119932 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.119948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.119958 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.222782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.222849 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.222863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.222885 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.222898 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.329682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.330040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.330185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.330319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.330445 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.433085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.433150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.433162 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.433181 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.433192 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.475786 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:41 crc kubenswrapper[4678]: E1206 10:37:41.476016 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.536448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.536516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.536531 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.536551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.536565 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.640130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.640737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.640916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.641123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.641325 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.752664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.752709 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.752719 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.752736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.752748 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.855403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.855438 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.855447 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.855463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.855472 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.957653 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.957687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.957699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.957716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.957729 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:41Z","lastTransitionTime":"2025-12-06T10:37:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.959023 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/0.log" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.959080 4678 generic.go:334] "Generic (PLEG): container finished" podID="388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c" containerID="34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87" exitCode=1 Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.959117 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerDied","Data":"34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87"} Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.959631 4678 scope.go:117] "RemoveContainer" containerID="34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.975062 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:41Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:41 crc kubenswrapper[4678]: I1206 10:37:41.987870 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:41Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.002134 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:41Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.016112 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.036022 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.046914 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.061366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.061418 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.061444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.061468 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.061482 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.066292 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.082664 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.103430 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.115351 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.127739 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.138213 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.148528 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.163760 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.163837 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.163850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.163870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.163887 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.166187 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.176238 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.185884 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.196845 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.213020 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.266307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.266341 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.266349 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.266364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.266373 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.369348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.369386 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.369399 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.369419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.369434 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.472735 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.472792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.472805 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.472829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.472843 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.475285 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.475313 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.475322 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:42 crc kubenswrapper[4678]: E1206 10:37:42.475418 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:42 crc kubenswrapper[4678]: E1206 10:37:42.475586 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:42 crc kubenswrapper[4678]: E1206 10:37:42.475706 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.575026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.575082 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.575095 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.575117 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.575129 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.678363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.678444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.678468 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.678556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.678584 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.781463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.781537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.781549 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.781570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.781585 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.884660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.884722 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.884734 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.884757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.884769 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.965121 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/0.log" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.966530 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerStarted","Data":"278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.981865 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.987477 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.987529 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.987541 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.987562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.987577 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:42Z","lastTransitionTime":"2025-12-06T10:37:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:42 crc kubenswrapper[4678]: I1206 10:37:42.997413 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:42Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.012607 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.025212 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.039464 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.051910 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.066971 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.090131 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.090201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.090246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.090257 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.090274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.090284 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.104075 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.114979 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.135531 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.149967 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.166993 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.186792 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.193508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.193564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.193579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.193596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.193608 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.196833 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.208864 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.220394 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.232062 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:43Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.295825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.295865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.295876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.295895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.295907 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.398655 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.398721 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.398742 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.398769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.398787 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.475558 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:43 crc kubenswrapper[4678]: E1206 10:37:43.475734 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.500851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.500996 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.501133 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.501230 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.501323 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.603814 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.603855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.603865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.603883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.603895 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.706513 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.706987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.707141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.707256 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.707373 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.809599 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.809645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.809657 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.809671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.809681 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.912294 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.912352 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.912364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.912389 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:43 crc kubenswrapper[4678]: I1206 10:37:43.912403 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:43Z","lastTransitionTime":"2025-12-06T10:37:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.017093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.017149 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.017164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.017187 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.017205 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.120110 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.120157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.120168 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.120186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.120199 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.222420 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.222456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.222466 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.222481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.222511 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.324967 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.325018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.325050 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.325070 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.325084 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.427260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.427289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.427298 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.427312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.427321 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.475860 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.475939 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:44 crc kubenswrapper[4678]: E1206 10:37:44.476003 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:44 crc kubenswrapper[4678]: E1206 10:37:44.476074 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.476171 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:44 crc kubenswrapper[4678]: E1206 10:37:44.476299 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.529471 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.529762 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.529829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.529894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.529956 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.632906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.632947 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.632959 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.632977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.632992 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.735751 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.736234 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.736428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.736699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.736855 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.839404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.839443 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.839451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.839469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.839480 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.941552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.941587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.941598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.941616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:44 crc kubenswrapper[4678]: I1206 10:37:44.941628 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:44Z","lastTransitionTime":"2025-12-06T10:37:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.043783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.043847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.043867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.043892 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.043910 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.146423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.146458 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.146469 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.146508 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.146520 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.248867 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.249357 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.249371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.249388 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.249399 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.352364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.352428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.352448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.352475 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.352522 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.467473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.467555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.467568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.467587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.467606 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.474930 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:45 crc kubenswrapper[4678]: E1206 10:37:45.476238 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.488283 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.500875 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.517509 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.530050 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.546394 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.565755 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.569570 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.569619 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.569628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.569644 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.569655 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.577452 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.586706 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.604686 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.614121 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.623970 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.633912 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.644785 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.657158 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.671974 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.673929 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.673992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.674010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.674032 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.674048 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.686458 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.705771 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.719191 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:45Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.777018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.777050 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.777061 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.777076 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.777086 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.879622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.879680 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.879693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.879711 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.879723 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.982742 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.982797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.982808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.982828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:45 crc kubenswrapper[4678]: I1206 10:37:45.982839 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:45Z","lastTransitionTime":"2025-12-06T10:37:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.085639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.085685 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.085694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.085712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.085723 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.188685 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.188751 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.188765 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.188791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.188810 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.291314 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.291359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.291368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.291390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.291402 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.394100 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.394143 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.394154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.394171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.394181 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.474904 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.475047 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:46 crc kubenswrapper[4678]: E1206 10:37:46.475243 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.475374 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:46 crc kubenswrapper[4678]: E1206 10:37:46.475714 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:46 crc kubenswrapper[4678]: E1206 10:37:46.475839 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.476042 4678 scope.go:117] "RemoveContainer" containerID="dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.498980 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.499037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.499050 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.499072 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.499091 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.601888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.601932 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.601941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.601954 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.601963 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.706832 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.707111 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.707122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.707179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.707193 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.811698 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.811727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.811737 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.811754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.811765 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.914211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.914241 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.914248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.914264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.914274 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:46Z","lastTransitionTime":"2025-12-06T10:37:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.980687 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/2.log" Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.988581 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7"} Dec 06 10:37:46 crc kubenswrapper[4678]: I1206 10:37:46.988951 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.013045 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.017046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.017093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.017106 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.017126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.017137 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.028001 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.041354 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.055531 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.072087 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.085402 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.097541 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.119533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.119574 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.119584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.119599 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.119610 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.121475 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.139035 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.158475 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.169824 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.185881 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.201778 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.214539 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.222063 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.222098 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.222109 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.222126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.222138 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.231837 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.248373 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.261929 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.273795 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:47Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.324664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.324714 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.324723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.324739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.324751 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.427789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.427846 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.427858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.427876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.427886 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.475585 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:47 crc kubenswrapper[4678]: E1206 10:37:47.475713 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.532247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.532313 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.532333 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.532359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.532377 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.635375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.635431 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.635450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.635474 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.635525 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.738008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.738063 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.738076 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.738096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.738122 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.840730 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.840834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.840856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.840917 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.840937 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.944250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.944287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.944297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.944312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.944322 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:47Z","lastTransitionTime":"2025-12-06T10:37:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.995843 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/3.log" Dec 06 10:37:47 crc kubenswrapper[4678]: I1206 10:37:47.996562 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/2.log" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.008398 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7" exitCode=1 Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.008648 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.008993 4678 scope.go:117] "RemoveContainer" containerID="dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.010067 4678 scope.go:117] "RemoveContainer" containerID="226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7" Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.010344 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.027409 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.047721 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.047777 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.047785 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.047803 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.047832 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.049292 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.067099 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.087334 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.103746 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.121628 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.131308 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.145694 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.150184 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.150331 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.150412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.150535 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.150613 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.158262 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.168379 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.182668 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.200820 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.215744 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.238666 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.253152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.253201 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.253216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.253236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.253250 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.261500 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dccd4db48253ae039f7490a8bf90a443e42ea53c37f400106dfef6410228ee3f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:20Z\\\",\\\"message\\\":\\\"c\\\\nI1206 10:37:20.401578 6198 ovn.go:134] Ensuring zone local for Pod openshift-kube-controller-manager/kube-controller-manager-crc in node crc\\\\nI1206 10:37:20.401585 6198 base_network_controller_pods.go:477] [default/openshift-multus/network-metrics-daemon-wfvj4] creating logical port openshift-multus_network-metrics-daemon-wfvj4 for pod on switch crc\\\\nI1206 10:37:20.401589 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc after 0 failed attempt(s)\\\\nI1206 10:37:20.401380 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401600 6198 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1206 10:37:20.401465 6198 obj_retry.go:303] Retry object setup: *v1.Pod openshift-ovn-kubernetes/ovnkube-node-5lfqt\\\\nI1206 10:37:20.401611 6198 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-fqd9d\\\\nI1206 10:37:20.401623 6198 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-fqd9d in node crc\\\\nI1206 10:37:20.401534 6198 default_network_controller.go:776] Recording success event on pod openshift-etcd/etcd-crc\\\\nI1206 10:37:20.401630 6198 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-fqd\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:47Z\\\",\\\"message\\\":\\\"{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 10:37:47.348881 6534 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.275429 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.285886 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.299058 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.355332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.355671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.355757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.355886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.355967 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.458723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.458808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.458828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.458855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.458875 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.474856 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.474968 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.475002 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.475158 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.475411 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.475555 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.562204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.562544 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.562618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.562705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.562797 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.578317 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.578391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.578409 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.578435 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.578454 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.594181 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.598052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.598189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.598276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.598369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.598457 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.612784 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.616677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.616719 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.616735 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.616759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.616777 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.631067 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.634690 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.634731 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.634743 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.634759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.634770 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.647781 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.651445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.651510 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.651528 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.651552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.651569 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.665074 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:48Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:48 crc kubenswrapper[4678]: E1206 10:37:48.665227 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.666997 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.667039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.667056 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.667078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.667095 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.769538 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.769587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.769603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.769628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.769647 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.872874 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.872917 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.872936 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.872963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.873020 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.976871 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.977359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.977561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.977716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:48 crc kubenswrapper[4678]: I1206 10:37:48.977847 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:48Z","lastTransitionTime":"2025-12-06T10:37:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.017089 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/3.log" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.023776 4678 scope.go:117] "RemoveContainer" containerID="226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7" Dec 06 10:37:49 crc kubenswrapper[4678]: E1206 10:37:49.024111 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.050180 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.073080 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.090199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.095473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.095536 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.095571 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.095595 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.097446 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.121367 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.140267 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.160746 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.176007 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.199775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.199816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.199833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.199859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.199876 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.200114 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.220533 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.233825 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.248298 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.272773 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.287397 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.302000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.302038 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.302051 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.302068 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.302081 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.305066 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.329051 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:47Z\\\",\\\"message\\\":\\\"{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 10:37:47.348881 6534 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.343027 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.361216 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.377460 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:49Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.404219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.404302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.404318 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.404337 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.404350 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.476053 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:49 crc kubenswrapper[4678]: E1206 10:37:49.476320 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.507723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.507770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.507780 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.507797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.507810 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.610387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.610973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.611175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.611369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.611648 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.714937 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.714995 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.715007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.715026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.715037 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.818404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.818466 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.818516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.818548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.818569 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.921851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.921894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.921905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.921920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:49 crc kubenswrapper[4678]: I1206 10:37:49.921932 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:49Z","lastTransitionTime":"2025-12-06T10:37:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.025229 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.025294 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.025305 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.025320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.025351 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.128816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.130103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.130404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.130707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.130960 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.234221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.234253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.234263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.234278 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.234288 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.337204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.337280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.337305 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.337339 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.337362 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.440610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.440664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.440681 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.440705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.440725 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.475937 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.475983 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:50 crc kubenswrapper[4678]: E1206 10:37:50.476194 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.476244 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:50 crc kubenswrapper[4678]: E1206 10:37:50.476389 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:50 crc kubenswrapper[4678]: E1206 10:37:50.476564 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.543777 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.543825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.543837 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.543854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.543870 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.648153 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.648232 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.648252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.648282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.648304 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.757481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.758089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.758219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.758335 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.758418 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.862072 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.862513 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.862587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.862659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.862727 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.966788 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.966865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.966886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.966913 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:50 crc kubenswrapper[4678]: I1206 10:37:50.966933 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:50Z","lastTransitionTime":"2025-12-06T10:37:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.070977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.071399 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.071517 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.071620 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.071715 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.175770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.175828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.175840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.175859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.175872 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.279664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.279743 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.279768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.279798 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.279822 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.382907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.382942 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.382951 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.382966 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.382976 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.475757 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:51 crc kubenswrapper[4678]: E1206 10:37:51.476035 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.486092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.486185 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.486206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.486235 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.486258 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.590422 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.590479 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.590535 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.590563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.590583 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.694244 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.694312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.694330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.694355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.694377 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.797513 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.797579 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.797592 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.797609 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.797622 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.901354 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.901426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.901445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.901472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:51 crc kubenswrapper[4678]: I1206 10:37:51.901524 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:51Z","lastTransitionTime":"2025-12-06T10:37:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.005760 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.005832 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.005855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.005884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.005908 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.109371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.109595 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.109631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.109660 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.109679 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.212058 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.212094 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.212106 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.212122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.212132 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.315326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.315387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.315408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.315437 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.315459 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.418717 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.418763 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.418773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.418786 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.418795 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.474985 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.475011 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.474984 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:52 crc kubenswrapper[4678]: E1206 10:37:52.475189 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:52 crc kubenswrapper[4678]: E1206 10:37:52.475338 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:52 crc kubenswrapper[4678]: E1206 10:37:52.475549 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.521584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.521669 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.521699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.521730 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.521749 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.625659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.625723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.625741 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.625767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.625785 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.728935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.729023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.729043 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.729074 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.729102 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.832452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.832970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.833160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.833342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.833466 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.936482 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.936605 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.936623 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.936651 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:52 crc kubenswrapper[4678]: I1206 10:37:52.936673 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:52Z","lastTransitionTime":"2025-12-06T10:37:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.041280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.041335 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.041353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.041378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.041395 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.144858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.144930 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.144963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.144992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.145013 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.247888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.247944 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.247960 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.247983 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.247998 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.350986 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.351036 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.351055 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.351078 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.351096 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.453387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.453436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.453444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.453460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.453470 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.475361 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:53 crc kubenswrapper[4678]: E1206 10:37:53.475539 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.561616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.561662 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.561671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.561984 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.562013 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.665822 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.665879 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.665890 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.665908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.665920 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.768927 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.768972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.768982 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.768997 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.769009 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.872178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.872245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.872262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.872285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.872303 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.975678 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.975777 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.975802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.975835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:53 crc kubenswrapper[4678]: I1206 10:37:53.975856 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:53Z","lastTransitionTime":"2025-12-06T10:37:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.078350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.078415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.078439 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.078462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.078480 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.203216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.203300 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.203323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.203353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.203375 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.307014 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.307193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.307217 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.307247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.307268 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.409923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.409978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.409993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.410011 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.410023 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.475749 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.475847 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:54 crc kubenswrapper[4678]: E1206 10:37:54.475914 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.475965 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:54 crc kubenswrapper[4678]: E1206 10:37:54.476025 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:54 crc kubenswrapper[4678]: E1206 10:37:54.476183 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.513062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.513384 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.513631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.513839 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.514039 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.617936 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.618000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.618016 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.618037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.618052 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.720754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.720814 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.720831 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.720858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.720876 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.823563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.823634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.823674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.823709 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.823735 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.927189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.927272 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.927290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.927317 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:54 crc kubenswrapper[4678]: I1206 10:37:54.927335 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:54Z","lastTransitionTime":"2025-12-06T10:37:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.030802 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.030851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.030864 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.030883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.030897 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.135118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.135190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.135213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.135242 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.135264 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.238756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.238804 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.238816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.238833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.238848 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.342773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.342828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.342839 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.342859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.342872 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.445583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.446004 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.446102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.446203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.446266 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.475341 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:55 crc kubenswrapper[4678]: E1206 10:37:55.476060 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.490168 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.504270 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.527936 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.542856 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.550826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.550895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.550906 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.550925 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.550944 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.558590 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.580658 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:47Z\\\",\\\"message\\\":\\\"{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 10:37:47.348881 6534 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.595533 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.617384 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.633886 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.654323 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.655475 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.655545 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.655557 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.655580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.655595 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.674059 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.687683 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.702093 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.718987 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.756429 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.760190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.760435 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.760450 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.760470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.760483 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.793641 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.819792 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.831734 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:55Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.863280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.863316 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.863326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.863341 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.863351 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.966026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.966094 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.966109 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.966128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:55 crc kubenswrapper[4678]: I1206 10:37:55.966140 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:55Z","lastTransitionTime":"2025-12-06T10:37:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.069093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.069179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.069204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.069273 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.069301 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.172400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.172464 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.172479 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.172519 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.172540 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.275674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.275718 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.275731 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.275747 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.275759 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.378323 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.378380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.378401 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.378430 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.378451 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.475203 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.475313 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:56 crc kubenswrapper[4678]: E1206 10:37:56.475405 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.475313 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:56 crc kubenswrapper[4678]: E1206 10:37:56.475611 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:56 crc kubenswrapper[4678]: E1206 10:37:56.475811 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.481286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.481320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.481328 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.481340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.481350 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.584276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.584354 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.584377 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.584412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.584436 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.687760 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.687818 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.687836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.687859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.687878 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.791114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.791183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.791205 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.791236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.791257 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.895622 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.895674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.895696 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.895723 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.895747 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.998688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.998843 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.998863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.998888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:56 crc kubenswrapper[4678]: I1206 10:37:56.998905 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:56Z","lastTransitionTime":"2025-12-06T10:37:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.101825 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.101896 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.101917 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.101944 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.101968 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.205580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.206179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.206204 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.206234 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.206254 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.309619 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.309674 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.309686 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.309704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.309716 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.413071 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.413128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.413145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.413170 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.413188 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.476152 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:57 crc kubenswrapper[4678]: E1206 10:37:57.476421 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.516271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.516359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.516377 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.516403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.516421 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.619865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.619943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.619965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.619998 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.620018 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.723873 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.723945 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.723963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.723984 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.723996 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.827680 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.827738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.827754 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.827775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.827789 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.932159 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.932208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.932230 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.932259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:57 crc kubenswrapper[4678]: I1206 10:37:57.932275 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:57Z","lastTransitionTime":"2025-12-06T10:37:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.034907 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.034965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.034976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.034995 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.035011 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.139057 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.139116 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.139134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.139160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.139178 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.209991 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.210181 4678 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.210294 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.210270102 +0000 UTC m=+147.053701551 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.241712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.241750 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.241762 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.241779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.241791 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.310970 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.311118 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.311153 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311181 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.311152677 +0000 UTC m=+147.154584126 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.311216 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311255 4678 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311304 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.31129325 +0000 UTC m=+147.154724689 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311343 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311344 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311377 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311393 4678 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311433 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.311421313 +0000 UTC m=+147.154852772 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311359 4678 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311458 4678 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.311526 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.311478625 +0000 UTC m=+147.154910064 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.344136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.344171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.344182 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.344195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.344205 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.446863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.446898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.446905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.446919 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.446929 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.475634 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.475708 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.475913 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.476015 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.476384 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.476910 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.550824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.550902 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.550924 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.550962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.550990 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.654689 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.655098 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.655189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.655331 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.655442 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.757826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.757866 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.757878 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.757894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.757904 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.784065 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.784111 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.784122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.784139 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.784152 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.797549 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.801102 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.801212 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.801231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.801259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.801278 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.814040 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.817419 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.817453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.817463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.817480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.817509 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.828493 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.832514 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.832550 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.832567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.832587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.832598 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.849871 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.855035 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.855172 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.855239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.855301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.855357 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.866915 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:37:58Z is after 2025-08-24T17:21:41Z" Dec 06 10:37:58 crc kubenswrapper[4678]: E1206 10:37:58.867445 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.869089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.869113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.869121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.869134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.869142 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.971576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.971986 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.972086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.972192 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:58 crc kubenswrapper[4678]: I1206 10:37:58.972289 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:58Z","lastTransitionTime":"2025-12-06T10:37:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.075432 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.075914 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.076081 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.076218 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.076324 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.179808 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.179858 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.179869 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.179894 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.179907 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.282753 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.283183 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.283449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.283710 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.283987 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.391052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.391148 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.392018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.392062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.392085 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.475827 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:37:59 crc kubenswrapper[4678]: E1206 10:37:59.476083 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.495290 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.495351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.495366 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.495387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.495406 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.598764 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.598820 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.598840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.598881 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.598910 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.701908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.701965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.701977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.701992 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.702001 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.805210 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.805275 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.805289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.805312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.805329 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.908318 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.908381 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.908404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.908436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:37:59 crc kubenswrapper[4678]: I1206 10:37:59.908459 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:37:59Z","lastTransitionTime":"2025-12-06T10:37:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.012640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.012733 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.012761 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.012793 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.012813 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.117062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.117136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.117154 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.117181 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.117199 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.220299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.220373 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.220394 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.220423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.220443 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.323342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.323401 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.323424 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.323446 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.323460 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.427189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.427231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.427244 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.427262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.427276 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.475307 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.475457 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:00 crc kubenswrapper[4678]: E1206 10:38:00.475476 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:00 crc kubenswrapper[4678]: E1206 10:38:00.475675 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.475718 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:00 crc kubenswrapper[4678]: E1206 10:38:00.475819 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.530989 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.531053 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.531068 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.531089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.531106 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.634141 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.634197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.634208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.634227 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.634261 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.737200 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.737265 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.737277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.737293 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.737307 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.840447 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.840562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.840582 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.840615 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.840643 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.945055 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.945128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.945146 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.945186 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:00 crc kubenswrapper[4678]: I1206 10:38:00.945225 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:00Z","lastTransitionTime":"2025-12-06T10:38:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.050193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.050252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.050271 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.050293 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.050310 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.153463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.153577 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.153600 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.153632 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.153655 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.257074 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.257121 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.257137 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.257160 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.257176 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.359967 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.360033 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.360057 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.360085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.360110 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.462701 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.462786 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.462816 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.462846 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.462866 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.475212 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:01 crc kubenswrapper[4678]: E1206 10:38:01.475406 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.566120 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.566180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.566203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.566226 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.566253 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.670103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.670170 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.670194 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.670222 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.670243 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.773676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.773725 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.773759 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.773779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.773791 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.877021 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.877126 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.877150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.877182 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.877202 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.980644 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.980687 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.980699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.980718 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:01 crc kubenswrapper[4678]: I1206 10:38:01.980731 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:01Z","lastTransitionTime":"2025-12-06T10:38:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.082855 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.082901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.082912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.082928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.082940 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.185801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.185856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.185865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.185886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.185910 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.288535 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.288574 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.288583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.288598 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.288609 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.391555 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.391623 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.391642 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.391669 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.391689 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.475713 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.475881 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.476153 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:02 crc kubenswrapper[4678]: E1206 10:38:02.476282 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:02 crc kubenswrapper[4678]: E1206 10:38:02.476441 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:02 crc kubenswrapper[4678]: E1206 10:38:02.476573 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.476610 4678 scope.go:117] "RemoveContainer" containerID="226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7" Dec 06 10:38:02 crc kubenswrapper[4678]: E1206 10:38:02.476850 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.493965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.494013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.494026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.494046 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.494059 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.597677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.597756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.597773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.597795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.597809 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.701120 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.701179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.701195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.701220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.701241 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.804925 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.804984 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.805007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.805039 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.805075 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.908253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.908338 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.908365 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.908398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:02 crc kubenswrapper[4678]: I1206 10:38:02.908419 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:02Z","lastTransitionTime":"2025-12-06T10:38:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.013247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.013325 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.013343 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.013373 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.013401 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.116572 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.116627 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.116644 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.116668 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.116688 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.220378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.220712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.220740 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.220771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.220796 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.323663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.323720 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.323736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.323757 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.323772 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.427403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.427479 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.427540 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.427568 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.427598 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.475538 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:03 crc kubenswrapper[4678]: E1206 10:38:03.475844 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.530776 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.530826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.530838 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.530854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.530867 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.634348 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.634413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.634431 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.634456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.634476 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.738030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.738096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.738109 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.738130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.738146 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.840792 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.840822 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.840832 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.840847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.840856 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.943944 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.944040 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.944057 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.944131 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:03 crc kubenswrapper[4678]: I1206 10:38:03.944151 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:03Z","lastTransitionTime":"2025-12-06T10:38:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.047220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.047302 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.047325 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.047353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.047371 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.149977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.150105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.150124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.150147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.150164 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.253675 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.253753 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.253769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.253790 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.253810 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.357171 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.357248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.357268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.357292 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.357311 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.461195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.461263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.461289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.461317 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.461338 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.475805 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.475859 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.475822 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:04 crc kubenswrapper[4678]: E1206 10:38:04.476184 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:04 crc kubenswrapper[4678]: E1206 10:38:04.476359 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:04 crc kubenswrapper[4678]: E1206 10:38:04.476580 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.564832 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.564935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.564964 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.564991 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.565012 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.668444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.669001 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.669052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.669085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.669201 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.771873 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.771927 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.771939 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.771959 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.771972 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.874700 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.874743 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.874752 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.874766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.874775 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.978295 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.978901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.979417 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.979805 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:04 crc kubenswrapper[4678]: I1206 10:38:04.979977 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:04Z","lastTransitionTime":"2025-12-06T10:38:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.083941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.084021 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.084043 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.084076 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.084095 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.187289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.187381 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.187407 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.187437 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.187456 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.291239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.291620 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.291714 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.291815 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.291908 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.395561 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.395989 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.396150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.396311 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.396347 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.475684 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:05 crc kubenswrapper[4678]: E1206 10:38:05.475850 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.498870 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.498941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.498963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.498995 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.499018 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.502206 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.521890 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.542103 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.561042 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.576900 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.589275 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.601114 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.601206 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.601228 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.601253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.601273 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.603510 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.617078 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.635876 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.653106 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.668176 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.681453 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.694866 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.704064 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.704303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.704396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.704511 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.704627 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.712751 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.727021 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.745653 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.764923 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.792651 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:47Z\\\",\\\"message\\\":\\\"{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 10:37:47.348881 6534 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:05Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.807124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.807383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.807604 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.807775 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.807876 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.910066 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.910145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.910163 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.910610 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:05 crc kubenswrapper[4678]: I1206 10:38:05.910743 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:05Z","lastTransitionTime":"2025-12-06T10:38:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.013908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.013968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.013983 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.014005 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.014018 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.117456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.117551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.117569 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.117594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.117612 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.220901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.220981 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.221017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.221052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.221072 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.324889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.324935 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.324951 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.324977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.324996 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.427191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.427614 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.427706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.427780 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.427849 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.475368 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.475682 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:06 crc kubenswrapper[4678]: E1206 10:38:06.475836 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.476098 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:06 crc kubenswrapper[4678]: E1206 10:38:06.476267 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:06 crc kubenswrapper[4678]: E1206 10:38:06.476403 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.530611 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.530656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.530675 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.530697 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.530711 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.634309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.634375 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.634392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.634422 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.634470 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.737088 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.737926 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.738015 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.738112 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.738202 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.840847 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.840903 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.840915 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.840939 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.840955 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.943926 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.944223 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.944289 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.944361 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:06 crc kubenswrapper[4678]: I1206 10:38:06.944424 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:06Z","lastTransitionTime":"2025-12-06T10:38:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.047322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.047412 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.047454 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.047515 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.047531 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.150912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.151263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.151362 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.151464 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.151560 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.254688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.254773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.254797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.254826 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.254849 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.357381 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.357452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.357470 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.357548 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.357567 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.461688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.462220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.462480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.462768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.462960 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.475340 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:07 crc kubenswrapper[4678]: E1206 10:38:07.475592 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.492808 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.566654 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.566743 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.566783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.566817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.566840 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.707772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.707851 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.707874 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.707903 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.707921 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.810157 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.810191 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.810202 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.810216 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.810226 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.912779 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.912822 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.912834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.912852 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:07 crc kubenswrapper[4678]: I1206 10:38:07.912863 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:07Z","lastTransitionTime":"2025-12-06T10:38:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.016007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.016072 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.016089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.016118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.016137 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.119123 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.119208 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.119233 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.119270 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.119294 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.222310 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.222358 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.222372 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.222391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.222407 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.326109 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.326165 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.326175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.326193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.326205 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.429663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.430195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.430681 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.431371 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.432234 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.475762 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.475843 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.475948 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:08 crc kubenswrapper[4678]: E1206 10:38:08.476750 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:08 crc kubenswrapper[4678]: E1206 10:38:08.476460 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:08 crc kubenswrapper[4678]: E1206 10:38:08.477012 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.535409 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.535465 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.535484 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.535540 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.535560 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.639386 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.639459 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.639478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.639536 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.639560 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.743175 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.743231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.743243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.743266 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.743284 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.846667 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.846805 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.846833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.846864 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.846884 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.950264 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.950324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.950335 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.950355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.950369 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.962017 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.962097 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.962118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.962147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.962172 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:08 crc kubenswrapper[4678]: E1206 10:38:08.985119 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:08Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.990893 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.990955 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.990978 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.991009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:08 crc kubenswrapper[4678]: I1206 10:38:08.991033 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:08Z","lastTransitionTime":"2025-12-06T10:38:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: E1206 10:38:09.012072 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:09Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.019963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.020025 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.020049 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.020089 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.020112 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: E1206 10:38:09.044280 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:09Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.049018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.049058 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.049073 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.049093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.049106 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: E1206 10:38:09.065198 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:09Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.070616 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.070658 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.070671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.070688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.070700 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: E1206 10:38:09.086838 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:09Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:09 crc kubenswrapper[4678]: E1206 10:38:09.087078 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.089147 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.089211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.089227 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.089245 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.089258 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.192136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.192219 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.192242 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.192274 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.192311 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.295432 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.295558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.295572 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.295593 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.295647 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.399398 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.399539 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.399563 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.399596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.399620 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.475958 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:09 crc kubenswrapper[4678]: E1206 10:38:09.476168 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.501968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.502008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.502026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.502048 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.502068 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.605307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.605810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.606023 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.606199 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.606365 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.710190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.710239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.710252 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.710269 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.710283 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.813834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.813965 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.813987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.814012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.814027 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.917649 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.917716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.917729 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.917749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:09 crc kubenswrapper[4678]: I1206 10:38:09.917762 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:09Z","lastTransitionTime":"2025-12-06T10:38:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.021193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.021250 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.021259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.021276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.021287 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.123897 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.124652 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.124760 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.124888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.124976 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.227694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.228092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.228189 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.228279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.228368 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.332190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.332236 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.332246 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.332263 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.332274 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.435193 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.435619 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.435707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.435782 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.435847 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.475176 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.475320 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:10 crc kubenswrapper[4678]: E1206 10:38:10.475587 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:10 crc kubenswrapper[4678]: E1206 10:38:10.475345 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.475606 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:10 crc kubenswrapper[4678]: E1206 10:38:10.475708 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.538575 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.538618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.538628 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.538643 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.538655 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.641129 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.641597 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.641776 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.641920 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.642048 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.744195 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.744553 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.744634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.744718 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.744783 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.847449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.847480 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.847515 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.847532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.847542 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.950916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.950958 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.950970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.950986 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:10 crc kubenswrapper[4678]: I1206 10:38:10.950997 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:10Z","lastTransitionTime":"2025-12-06T10:38:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.053789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.053854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.053866 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.053905 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.053922 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.162331 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.162414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.162435 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.162463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.162514 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.265552 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.265585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.265594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.265607 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.265673 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.368948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.369294 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.369595 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.369707 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.369785 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.472445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.472503 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.472515 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.472546 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.472559 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.475842 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:11 crc kubenswrapper[4678]: E1206 10:38:11.476070 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.575164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.575625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.575770 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.575888 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.575976 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.678384 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.678431 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.678441 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.678457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.678469 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.781532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.781575 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.781588 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.781608 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.781620 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.884625 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.884667 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.884681 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.884696 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.884708 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.988018 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.988069 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.988082 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.988104 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:11 crc kubenswrapper[4678]: I1206 10:38:11.988119 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:11Z","lastTransitionTime":"2025-12-06T10:38:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.091052 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.091641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.092103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.092280 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.092404 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.195161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.195213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.195230 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.195251 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.195270 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.297639 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.297677 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.297688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.297704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.297713 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.400993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.401073 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.401093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.401127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.401147 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.475225 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.475265 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.475329 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:12 crc kubenswrapper[4678]: E1206 10:38:12.475379 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:12 crc kubenswrapper[4678]: E1206 10:38:12.475539 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:12 crc kubenswrapper[4678]: E1206 10:38:12.475609 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.504359 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.504432 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.504444 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.504462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.504473 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.607726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.607794 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.607813 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.607833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.607848 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.711443 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.711562 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.711585 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.711613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.711635 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.780953 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:12 crc kubenswrapper[4678]: E1206 10:38:12.781239 4678 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:38:12 crc kubenswrapper[4678]: E1206 10:38:12.781325 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs podName:5fdea9d8-255a-4ec7-8095-4489828f4136 nodeName:}" failed. No retries permitted until 2025-12-06 10:39:16.781298029 +0000 UTC m=+161.624729508 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs") pod "network-metrics-daemon-wfvj4" (UID: "5fdea9d8-255a-4ec7-8095-4489828f4136") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.815178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.815316 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.815341 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.815369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.815390 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.918798 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.919363 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.919861 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.920210 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:12 crc kubenswrapper[4678]: I1206 10:38:12.920632 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:12Z","lastTransitionTime":"2025-12-06T10:38:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.023732 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.023833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.023856 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.023887 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.023907 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.126646 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.127103 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.127242 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.127400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.127638 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.230715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.230769 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.230783 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.230801 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.230813 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.333351 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.333403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.333427 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.333449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.333463 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.435898 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.435940 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.435950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.435967 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.435978 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.474934 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:13 crc kubenswrapper[4678]: E1206 10:38:13.475087 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.475901 4678 scope.go:117] "RemoveContainer" containerID="226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7" Dec 06 10:38:13 crc kubenswrapper[4678]: E1206 10:38:13.476227 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\"" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.538972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.539044 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.539062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.539086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.539110 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.642796 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.642853 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.642863 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.642882 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.642896 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.746547 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.746617 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.746637 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.746664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.746683 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.850258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.850331 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.850355 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.850379 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.850398 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.954030 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.954122 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.954144 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.954176 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:13 crc kubenswrapper[4678]: I1206 10:38:13.954199 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:13Z","lastTransitionTime":"2025-12-06T10:38:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.057426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.057483 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.057533 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.057559 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.057576 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.160426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.160471 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.160482 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.160521 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.160533 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.263086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.263162 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.263187 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.263220 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.263241 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.366150 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.366730 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.366923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.367372 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.367667 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.470403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.470452 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.470462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.470478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.470507 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.475649 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.475727 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:14 crc kubenswrapper[4678]: E1206 10:38:14.475760 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.475649 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:14 crc kubenswrapper[4678]: E1206 10:38:14.475828 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:14 crc kubenswrapper[4678]: E1206 10:38:14.475880 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.573416 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.573475 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.573524 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.573554 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.573573 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.676197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.676294 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.676320 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.676350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.676372 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.778857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.778913 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.778928 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.778950 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.778968 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.882676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.882760 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.882772 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.882795 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.882808 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.986282 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.986330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.986340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.986364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:14 crc kubenswrapper[4678]: I1206 10:38:14.986376 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:14Z","lastTransitionTime":"2025-12-06T10:38:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.089404 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.089445 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.089457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.089475 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.089504 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.192286 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.192329 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.192340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.192357 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.192369 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.295564 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.295667 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.295686 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.295712 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.295730 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.398878 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.398977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.399002 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.399037 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.399065 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.475962 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:15 crc kubenswrapper[4678]: E1206 10:38:15.476177 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.495908 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.501596 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.501641 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.501659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.501683 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.501702 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.512551 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.523870 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.536210 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.548888 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.573636 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.584989 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cd69eef-734b-450f-932a-f929195ae285\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b15300e06c9969cb618aa54a705101e37282d0992fd20ed8bb251bee3ef297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69956e54a4d82ac7a8c7d6e7bce9bb238e0d4d13d857ed8de5aa2dbf6980ff58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69956e54a4d82ac7a8c7d6e7bce9bb238e0d4d13d857ed8de5aa2dbf6980ff58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.603542 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.605602 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.605663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.605673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.605694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.605705 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.616820 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.633467 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.652320 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:47Z\\\",\\\"message\\\":\\\"{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 10:37:47.348881 6534 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.665715 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.676978 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.690031 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.701866 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.708086 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.708116 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.708124 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.708138 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.708149 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.717318 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.731307 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.745687 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.760336 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:15Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.811032 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.811067 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.811077 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.811093 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.811111 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.914629 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.914711 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.914734 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.914767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:15 crc kubenswrapper[4678]: I1206 10:38:15.914791 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:15Z","lastTransitionTime":"2025-12-06T10:38:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.017720 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.017773 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.017788 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.017806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.017820 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.120843 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.120876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.120884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.120899 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.120909 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.223745 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.223819 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.223841 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.223868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.223887 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.327306 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.327369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.327382 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.327403 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.327419 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.430911 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.430986 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.431012 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.431049 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.431072 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.475451 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.475554 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.475450 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:16 crc kubenswrapper[4678]: E1206 10:38:16.475757 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:16 crc kubenswrapper[4678]: E1206 10:38:16.475876 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:16 crc kubenswrapper[4678]: E1206 10:38:16.476036 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.534133 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.534179 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.534197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.534215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.534227 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.636387 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.636440 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.636453 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.636472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.636518 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.739634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.739704 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.739730 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.739776 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.739801 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.843642 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.843728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.843744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.843766 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.843782 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.947028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.947094 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.947120 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.947152 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:16 crc kubenswrapper[4678]: I1206 10:38:16.947175 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:16Z","lastTransitionTime":"2025-12-06T10:38:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.051852 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.051909 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.051927 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.051952 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.051969 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.155178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.155253 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.155262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.155276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.155289 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.259440 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.259507 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.259517 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.259535 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.259548 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.363070 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.363106 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.363118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.363133 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.363146 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.466423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.466473 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.466484 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.466537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.466549 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.475856 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:17 crc kubenswrapper[4678]: E1206 10:38:17.476001 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.569346 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.569385 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.569397 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.569414 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.569425 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.672542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.672699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.672713 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.672746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.672758 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.776370 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.776411 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.776423 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.776440 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.776451 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.879410 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.879476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.879523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.879551 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.879570 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.983279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.983330 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.983340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.983367 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:17 crc kubenswrapper[4678]: I1206 10:38:17.983380 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:17Z","lastTransitionTime":"2025-12-06T10:38:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.086258 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.086312 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.086327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.086350 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.086363 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.190391 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.191474 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.191676 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.191834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.191967 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.294936 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.294977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.294987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.295010 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.295023 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.398989 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.399663 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.399714 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.399744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.399767 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.475251 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.475326 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.476005 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:18 crc kubenswrapper[4678]: E1206 10:38:18.476231 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:18 crc kubenswrapper[4678]: E1206 10:38:18.476480 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:18 crc kubenswrapper[4678]: E1206 10:38:18.477021 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.503319 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.503853 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.504043 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.504248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.504452 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.607609 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.607668 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.607679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.607694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.607705 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.710908 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.710961 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.710972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.710990 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.711001 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.813940 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.814368 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.814618 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.814844 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.815021 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.918472 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.918523 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.918532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.918547 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:18 crc kubenswrapper[4678]: I1206 10:38:18.918557 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:18Z","lastTransitionTime":"2025-12-06T10:38:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.022332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.022410 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.022430 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.022457 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.022476 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.125516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.125556 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.125567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.125582 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.125592 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.228748 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.228786 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.228796 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.228812 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.228823 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.300436 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.300545 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.300567 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.300594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.300614 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: E1206 10:38:19.318129 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.329749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.329812 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.329833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.329859 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.329879 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: E1206 10:38:19.350245 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.355304 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.355345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.355356 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.355376 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.355389 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: E1206 10:38:19.375186 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.380026 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.380092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.380110 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.380135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.380154 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: E1206 10:38:19.396824 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.401362 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.401413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.401426 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.401449 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.401464 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: E1206 10:38:19.417445 4678 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148064Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608864Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-06T10:38:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"d036a3e2-4349-4778-85ff-e71cad7388b7\\\",\\\"systemUUID\\\":\\\"05fdc3ed-7324-40a6-b64a-9e2883945f0d\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:19Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:19 crc kubenswrapper[4678]: E1206 10:38:19.417617 4678 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.419656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.419702 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.419716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.419735 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.419751 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.475271 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:19 crc kubenswrapper[4678]: E1206 10:38:19.475454 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.522085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.522162 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.522198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.522215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.522245 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.625682 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.625738 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.625749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.625767 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.625779 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.728715 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.728789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.728806 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.728836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.728853 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.832448 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.832594 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.832624 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.832659 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.832685 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.936408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.936478 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.936532 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.936560 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:19 crc kubenswrapper[4678]: I1206 10:38:19.936580 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:19Z","lastTransitionTime":"2025-12-06T10:38:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.040239 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.040297 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.040308 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.040327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.040337 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.143650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.143693 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.143705 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.143744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.143758 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.247603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.247681 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.247706 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.247744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.247768 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.350829 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.350923 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.350945 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.350973 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.350993 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.453739 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.453800 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.453813 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.453835 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.453847 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.475360 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.475360 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.475605 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:20 crc kubenswrapper[4678]: E1206 10:38:20.475736 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:20 crc kubenswrapper[4678]: E1206 10:38:20.475846 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:20 crc kubenswrapper[4678]: E1206 10:38:20.476028 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.556516 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.556576 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.556587 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.556606 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.556620 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.660170 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.660231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.660242 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.660262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.660276 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.763255 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.763307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.763321 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.763344 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.763358 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.865883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.865931 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.865943 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.865963 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.865976 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.969105 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.969164 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.969180 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.969197 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:20 crc kubenswrapper[4678]: I1206 10:38:20.969207 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:20Z","lastTransitionTime":"2025-12-06T10:38:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.071238 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.071277 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.071285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.071299 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.071309 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.174768 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.174836 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.174845 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.174864 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.174875 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.277027 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.277092 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.277110 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.277134 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.277151 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.380460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.380665 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.380680 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.380703 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.380715 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.475415 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:21 crc kubenswrapper[4678]: E1206 10:38:21.475659 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.484085 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.484182 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.484210 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.484248 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.484276 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.587213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.587279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.587296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.587321 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.587336 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.690601 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.690697 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.690720 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.690749 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.690768 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.793834 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.793912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.793938 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.793969 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.793998 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.896336 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.896401 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.896415 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.896437 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:21 crc kubenswrapper[4678]: I1206 10:38:21.896453 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:21Z","lastTransitionTime":"2025-12-06T10:38:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.000166 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.000231 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.000254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.000285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.000308 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.103876 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.103970 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.103988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.104062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.104082 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.207949 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.208028 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.208047 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.208075 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.208095 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.311327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.311380 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.311392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.311413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.311429 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.414353 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.414386 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.414396 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.414413 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.414421 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.475318 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.475341 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:22 crc kubenswrapper[4678]: E1206 10:38:22.475669 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.475857 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:22 crc kubenswrapper[4678]: E1206 10:38:22.475838 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:22 crc kubenswrapper[4678]: E1206 10:38:22.476021 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.517962 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.518009 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.518025 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.518049 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.518067 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.621303 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.621354 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.621369 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.621389 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.621404 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.723884 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.723939 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.723953 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.723977 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.723993 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.827128 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.827178 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.827190 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.827211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.827226 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.929941 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.929983 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.929993 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.930016 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:22 crc kubenswrapper[4678]: I1206 10:38:22.930027 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:22Z","lastTransitionTime":"2025-12-06T10:38:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.032887 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.033135 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.033200 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.033321 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.033401 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.135784 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.136108 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.136188 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.136251 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.136311 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.239979 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.240067 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.240094 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.240125 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.240150 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.343821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.343883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.343895 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.343916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.343934 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.447688 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.447755 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.447778 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.447810 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.447832 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.475934 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:23 crc kubenswrapper[4678]: E1206 10:38:23.476173 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.550854 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.550911 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.550929 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.550954 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.550977 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.654771 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.654817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.654828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.654846 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.654859 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.757549 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.757627 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.757650 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.757684 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.757706 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.860377 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.860463 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.860520 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.860559 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.860587 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.965205 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.965307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.965332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.965364 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:23 crc kubenswrapper[4678]: I1206 10:38:23.965393 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:23Z","lastTransitionTime":"2025-12-06T10:38:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.068307 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.068383 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.068400 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.068428 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.068447 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.171507 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.171583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.171603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.171631 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.171649 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.275288 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.275372 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.275392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.275417 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.275435 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.378233 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.378311 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.378326 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.378342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.378353 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.475792 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.475792 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.475797 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:24 crc kubenswrapper[4678]: E1206 10:38:24.476079 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:24 crc kubenswrapper[4678]: E1206 10:38:24.476181 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:24 crc kubenswrapper[4678]: E1206 10:38:24.476378 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.480995 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.481049 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.481068 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.481091 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.481296 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.584196 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.584234 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.584243 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.584260 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.584270 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.687812 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.687857 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.687869 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.687886 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.687899 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.790776 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.790833 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.790850 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.790875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.790893 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.894062 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.894107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.894118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.894136 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.894148 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.996791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.997275 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.997443 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.997640 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:24 crc kubenswrapper[4678]: I1206 10:38:24.997765 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:24Z","lastTransitionTime":"2025-12-06T10:38:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.100812 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.101182 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.101360 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.101513 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.101634 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.204161 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.204589 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.204736 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.204890 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.205046 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.308360 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.308774 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.309434 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.309584 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.309694 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.412638 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.412672 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.412680 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.412699 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.412711 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.476072 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:25 crc kubenswrapper[4678]: E1206 10:38:25.477126 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.488427 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a6b30be2-60a6-4733-be36-3f70c6ca6b03\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://65db652b41908d10d720c6416534e97a388866da41c23761b943e319787fd56f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pxkkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-bkhrd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.499447 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7cd69eef-734b-450f-932a-f929195ae285\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5b15300e06c9969cb618aa54a705101e37282d0992fd20ed8bb251bee3ef297\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://69956e54a4d82ac7a8c7d6e7bce9bb238e0d4d13d857ed8de5aa2dbf6980ff58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://69956e54a4d82ac7a8c7d6e7bce9bb238e0d4d13d857ed8de5aa2dbf6980ff58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.516840 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.517095 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.517203 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.517281 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.517360 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.519984 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc57de55-2af5-4914-8c50-0f8c14549243\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7c999cbbc6348e10961f7fe3adcf41eaccf0fcf2a09012699929907c969a38c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ffa0a2c78891bbe0453bc7b2c1f9e82114dd368d937c62f9946a257e8bb0a7a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eeef344c17008ebe0eec983be611bf2adf09aaafa8b3c16ee9c36f4cfce6d4a7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a50da283e61a3ba0e0d7757b07444f83349ce52c8257f5f3394a715be7807688\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb530b0720bcc3b14001e9ad421d23093a080b005a94b82cc2bfc45160f0579e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://489ae8cdd4913a593a396f327fa7a5332fac6b59b302a0614c541e1c13eded38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://50b9859b52bd44442ad7190762962beb0301730e0225a7a4d211a95c2c8ede40\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://64c97e730b32d74c3980bea8874ec3a931e39e005aa94f7dff4003b311b16185\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.535935 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.550998 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3966f3338abdecc3a2b971d505048f148bca1f09a826321b028bac6e06172b57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.563052 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-fqd9d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e219fa13-5c09-41c9-b89f-c00cd49c4d52\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://36e1b348d6b6c071f45580123f210882bc480d1a9ae7cb1076d55966126e2802\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dktr6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-fqd9d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.574585 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-dvc6z" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3daabe49-ee26-4bc3-acef-3d376a17fcd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://094f4bfcdf942435b434cc67bfcb51b24e030c7f48f00d7e34823662c47efc2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tjcnv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:56Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-dvc6z\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.586939 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c878da-974b-44a1-b97c-1c5590fa965e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53b03307dcb3812ce9151613b91e85875b4227cd41a0fbb2c42583ab4eb320a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d58465198a10dcbe81dcedd596d4ef803b33007a8a05b79507c5997d36aac789\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://375b26c93fbc3cdbf133a44efea98cff62f2d044b4cdf1a904838badb369f20e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1acf1392ccc44193f20a098e423ea694887cb0a12169ee1726a84ce84d1c1539\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.601157 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.615064 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce2f1906e83c9801feb10c69b240a12484a24bcb21cb16a6a8933016fcf8e2e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7a9a5d3a3f9e995a3493918418640c29313d0de69ee7e7d404a39d19c8e73ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.619597 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.619656 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.619670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.619698 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.619716 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.627797 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.646232 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5cca846-f84b-4924-b1c6-4ec6cea71a65\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:47Z\\\",\\\"message\\\":\\\"{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-api/machine-api-operator-machine-webhook_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-api/machine-api-operator-machine-webhook\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.250\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1206 10:37:47.348881 6534 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: fa\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:37:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-5lfqt_openshift-ovn-kubernetes(c5cca846-f84b-4924-b1c6-4ec6cea71a65)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qjjj4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-5lfqt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.661914 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fc07bde-92a6-4589-b434-ac369f244272\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-06T10:36:53Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1206 10:36:47.894915 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1206 10:36:47.896741 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1703256061/tls.crt::/tmp/serving-cert-1703256061/tls.key\\\\\\\"\\\\nI1206 10:36:53.423214 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1206 10:36:53.435860 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1206 10:36:53.435890 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1206 10:36:53.435946 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1206 10:36:53.435955 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1206 10:36:53.450929 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1206 10:36:53.450991 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1206 10:36:53.451519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451533 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1206 10:36:53.451540 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1206 10:36:53.451544 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1206 10:36:53.451548 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1206 10:36:53.451553 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1206 10:36:53.455103 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.678618 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c96c711a61ef2cc3693c8e7679afa42eaf7c36a9e422c21ba45ad4a52ae02c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.696398 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c41b983-e061-4cf2-94aa-b0926999502e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5707886499e3e4157efa2f5e1668bd5b97596de934ff6cee5964e11ee4c41d95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://196ccec6206ba2a8a0a8b45d5eecf3bfe99e8ff6662025f0826d2565ed34c614\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://168ae00ac58bb3f0cce77f6cffadbb4eb0f0a5c8e52f8e0a92386eca366e661c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:56Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c02fff80e4a418ee387b0c7a0ef1c8ffb7a28fe6cfe6a8ce7d59818876e7dd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ac5a534b3425a134bcb3a817fda3618891f6fdb2edeacfac90fd21d1cb6b792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1bcb9a80343a81368cfeb09bdebd753fd546004c7c938be707e36dd1c2c7253f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:36:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ac4146e0e7a74b9390d8c7c11a631264aab894c39f65b8e4b8f7825a78dc62eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-06T10:37:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kqblr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qzmlf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.710232 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-z5s25" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-06T10:37:41Z\\\",\\\"message\\\":\\\"2025-12-06T10:36:55+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2\\\\n2025-12-06T10:36:55+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_7798fd05-f949-4313-806c-6539a1888aa2 to /host/opt/cni/bin/\\\\n2025-12-06T10:36:55Z [verbose] multus-daemon started\\\\n2025-12-06T10:36:55Z [verbose] Readiness Indicator file check\\\\n2025-12-06T10:37:40Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-06T10:36:55Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dzbth\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:54Z\\\"}}\" for pod \"openshift-multus\"/\"multus-z5s25\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.722022 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.722327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.722464 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.722583 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.722665 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.727443 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b0811b3-b21b-475f-ac97-66a7934fd021\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:36:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56b839399cf880371d3ab1ec5a7f393a3e8d5892332e753f6f94fa3144305f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39be5407e47cf6e56c6663da18eb096cbebf91ba239615ed46571565fd4e1725\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://187c09c35ddbb9fed04b1e28fa985ae533b3c76571e99070e66aab60d21c4b5e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:36:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:36:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.740578 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25d9d216-9317-4d16-a1b6-85cd6d1e2543\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dfea0a3c5e9c114fd902814dd00cb83330030961766da0413250b0dab9a2a28\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f990f8da91136ad6fe1a122d52acce1bca3a9c7e1b01ee81d6d98809d936839\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-06T10:37:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rb9ns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:06Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-p9cql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.755169 4678 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fdea9d8-255a-4ec7-8095-4489828f4136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-06T10:37:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lcbs6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-06T10:37:08Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-wfvj4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-06T10:38:25Z is after 2025-08-24T17:21:41Z" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.825756 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.825797 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.825807 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.825824 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.825836 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.929671 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.929734 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.929745 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.929764 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:25 crc kubenswrapper[4678]: I1206 10:38:25.929777 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:25Z","lastTransitionTime":"2025-12-06T10:38:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.037476 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.037544 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.037558 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.037580 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.037601 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.141613 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.142211 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.142456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.142728 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.142910 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.246388 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.246451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.246462 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.246481 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.246516 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.350247 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.350296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.350309 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.350332 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.350348 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.454213 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.454268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.454279 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.454296 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.454309 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.475882 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.475929 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:26 crc kubenswrapper[4678]: E1206 10:38:26.476071 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.476133 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:26 crc kubenswrapper[4678]: E1206 10:38:26.476342 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:26 crc kubenswrapper[4678]: E1206 10:38:26.476692 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.556781 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.556848 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.556868 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.556896 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.556921 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.660007 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.660113 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.660130 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.660148 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.660160 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.763227 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.763287 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.763301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.763324 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.763339 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.866305 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.866378 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.866392 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.866416 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.866433 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.969933 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.970000 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.970096 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.970125 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:26 crc kubenswrapper[4678]: I1206 10:38:26.970184 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:26Z","lastTransitionTime":"2025-12-06T10:38:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.073914 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.073968 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.073987 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.074013 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.074032 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.177732 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.177799 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.177821 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.177842 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.177858 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.281020 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.281390 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.281664 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.281875 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.282024 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.386129 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.386679 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.386865 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.387033 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.387197 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.475357 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:27 crc kubenswrapper[4678]: E1206 10:38:27.475594 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.490482 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.490565 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.490582 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.490603 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.490622 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.595789 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.595860 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.595879 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.595916 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.595940 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.700008 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.700084 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.700107 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.700143 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.700172 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.803342 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.803431 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.803456 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.803521 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.803547 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.906846 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.906889 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.906901 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.906922 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:27 crc kubenswrapper[4678]: I1206 10:38:27.906937 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:27Z","lastTransitionTime":"2025-12-06T10:38:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.009198 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.009262 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.009275 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.009298 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.009310 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.112537 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.112632 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.112645 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.112673 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.112686 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.186362 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/1.log" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.187107 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/0.log" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.187161 4678 generic.go:334] "Generic (PLEG): container finished" podID="388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c" containerID="278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13" exitCode=1 Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.187201 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerDied","Data":"278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.187242 4678 scope.go:117] "RemoveContainer" containerID="34cee54f10d0467ef0828f7903552bc6e4e32e76da28870108a496252d702f87" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.188477 4678 scope.go:117] "RemoveContainer" containerID="278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13" Dec 06 10:38:28 crc kubenswrapper[4678]: E1206 10:38:28.189106 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-z5s25_openshift-multus(388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c)\"" pod="openshift-multus/multus-z5s25" podUID="388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.217254 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.217310 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.217322 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.217345 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.217368 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.235805 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-fqd9d" podStartSLOduration=95.23577068 podStartE2EDuration="1m35.23577068s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.220865573 +0000 UTC m=+113.064297042" watchObservedRunningTime="2025-12-06 10:38:28.23577068 +0000 UTC m=+113.079202159" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.253521 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podStartSLOduration=95.253474214 podStartE2EDuration="1m35.253474214s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.236869616 +0000 UTC m=+113.080301075" watchObservedRunningTime="2025-12-06 10:38:28.253474214 +0000 UTC m=+113.096905663" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.253874 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.253868483 podStartE2EDuration="21.253868483s" podCreationTimestamp="2025-12-06 10:38:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.253358521 +0000 UTC m=+113.096789970" watchObservedRunningTime="2025-12-06 10:38:28.253868483 +0000 UTC m=+113.097299932" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.305037 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=92.305010696 podStartE2EDuration="1m32.305010696s" podCreationTimestamp="2025-12-06 10:36:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.298349161 +0000 UTC m=+113.141780640" watchObservedRunningTime="2025-12-06 10:38:28.305010696 +0000 UTC m=+113.148442165" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.319727 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.319791 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.319803 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.319817 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.319831 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.380471 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-dvc6z" podStartSLOduration=95.380448057 podStartE2EDuration="1m35.380448057s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.380105189 +0000 UTC m=+113.223536648" watchObservedRunningTime="2025-12-06 10:38:28.380448057 +0000 UTC m=+113.223879536" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.401959 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=66.401935858 podStartE2EDuration="1m6.401935858s" podCreationTimestamp="2025-12-06 10:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.400933955 +0000 UTC m=+113.244365404" watchObservedRunningTime="2025-12-06 10:38:28.401935858 +0000 UTC m=+113.245367297" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.422285 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.422327 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.422340 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.422362 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.422374 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.475635 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.475726 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.475809 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:28 crc kubenswrapper[4678]: E1206 10:38:28.476596 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:28 crc kubenswrapper[4678]: E1206 10:38:28.476736 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.476825 4678 scope.go:117] "RemoveContainer" containerID="226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7" Dec 06 10:38:28 crc kubenswrapper[4678]: E1206 10:38:28.476915 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.481751 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=95.48173246 podStartE2EDuration="1m35.48173246s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.46411556 +0000 UTC m=+113.307546999" watchObservedRunningTime="2025-12-06 10:38:28.48173246 +0000 UTC m=+113.325163899" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.513998 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qzmlf" podStartSLOduration=95.513976993 podStartE2EDuration="1m35.513976993s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.513139863 +0000 UTC m=+113.356571302" watchObservedRunningTime="2025-12-06 10:38:28.513976993 +0000 UTC m=+113.357408432" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.525215 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.525259 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.525268 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.525284 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.525296 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.545863 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=94.545820556 podStartE2EDuration="1m34.545820556s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.54554432 +0000 UTC m=+113.388975749" watchObservedRunningTime="2025-12-06 10:38:28.545820556 +0000 UTC m=+113.389252035" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.565925 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-p9cql" podStartSLOduration=94.565901234 podStartE2EDuration="1m34.565901234s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:28.565593587 +0000 UTC m=+113.409025066" watchObservedRunningTime="2025-12-06 10:38:28.565901234 +0000 UTC m=+113.409332673" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.628843 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.628900 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.628912 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.628933 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.628952 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.731408 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.731460 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.731474 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.731828 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.731860 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.842145 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.842221 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.842276 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.842301 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.842315 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.944634 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.944670 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.944678 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.944691 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:28 crc kubenswrapper[4678]: I1206 10:38:28.944701 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:28Z","lastTransitionTime":"2025-12-06T10:38:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.047927 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.047976 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.047988 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.048006 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.048018 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:29Z","lastTransitionTime":"2025-12-06T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.150451 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.150510 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.150542 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.150560 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.150572 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:29Z","lastTransitionTime":"2025-12-06T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.193252 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/1.log" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.195837 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/3.log" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.199024 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerStarted","Data":"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.199574 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.227365 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podStartSLOduration=96.227345311 podStartE2EDuration="1m36.227345311s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:29.227095125 +0000 UTC m=+114.070526564" watchObservedRunningTime="2025-12-06 10:38:29.227345311 +0000 UTC m=+114.070776770" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.253694 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.253734 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.253744 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.253761 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.253772 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:29Z","lastTransitionTime":"2025-12-06T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.324230 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-wfvj4"] Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.324376 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:29 crc kubenswrapper[4678]: E1206 10:38:29.324473 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.356883 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.356937 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.356948 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.356972 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.356992 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:29Z","lastTransitionTime":"2025-12-06T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.445076 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.445118 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.445127 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.445142 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.445153 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:29Z","lastTransitionTime":"2025-12-06T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.459669 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.459716 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.459726 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.459746 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.459761 4678 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-06T10:38:29Z","lastTransitionTime":"2025-12-06T10:38:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.475411 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:29 crc kubenswrapper[4678]: E1206 10:38:29.475631 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.495629 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7"] Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.496753 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.499315 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.499432 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.499327 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.499693 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.583897 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/54054d9a-8094-416f-bb9e-6efaae9c561d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.584052 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/54054d9a-8094-416f-bb9e-6efaae9c561d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.584140 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54054d9a-8094-416f-bb9e-6efaae9c561d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.584179 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/54054d9a-8094-416f-bb9e-6efaae9c561d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.584350 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54054d9a-8094-416f-bb9e-6efaae9c561d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.685667 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54054d9a-8094-416f-bb9e-6efaae9c561d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.685752 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/54054d9a-8094-416f-bb9e-6efaae9c561d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.685781 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/54054d9a-8094-416f-bb9e-6efaae9c561d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.685809 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54054d9a-8094-416f-bb9e-6efaae9c561d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.686015 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/54054d9a-8094-416f-bb9e-6efaae9c561d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.685878 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/54054d9a-8094-416f-bb9e-6efaae9c561d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.686327 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/54054d9a-8094-416f-bb9e-6efaae9c561d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.686646 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/54054d9a-8094-416f-bb9e-6efaae9c561d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.697236 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54054d9a-8094-416f-bb9e-6efaae9c561d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.705894 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54054d9a-8094-416f-bb9e-6efaae9c561d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x7hv7\" (UID: \"54054d9a-8094-416f-bb9e-6efaae9c561d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:29 crc kubenswrapper[4678]: I1206 10:38:29.817636 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" Dec 06 10:38:30 crc kubenswrapper[4678]: I1206 10:38:30.203688 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" event={"ID":"54054d9a-8094-416f-bb9e-6efaae9c561d","Type":"ContainerStarted","Data":"52c92e726c58e08ad56cc7344b977b1c7fd43ed8d1b920efa18d204dad8ceae6"} Dec 06 10:38:30 crc kubenswrapper[4678]: I1206 10:38:30.474953 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:30 crc kubenswrapper[4678]: E1206 10:38:30.475171 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:30 crc kubenswrapper[4678]: I1206 10:38:30.474979 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:30 crc kubenswrapper[4678]: E1206 10:38:30.475624 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:31 crc kubenswrapper[4678]: I1206 10:38:31.209836 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" event={"ID":"54054d9a-8094-416f-bb9e-6efaae9c561d","Type":"ContainerStarted","Data":"816ac46837b33076900e42bad435f30d4bec64075ca37bb53d53348483fecf2f"} Dec 06 10:38:31 crc kubenswrapper[4678]: I1206 10:38:31.239138 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x7hv7" podStartSLOduration=98.23910425 podStartE2EDuration="1m38.23910425s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:31.236298204 +0000 UTC m=+116.079729693" watchObservedRunningTime="2025-12-06 10:38:31.23910425 +0000 UTC m=+116.082535759" Dec 06 10:38:31 crc kubenswrapper[4678]: I1206 10:38:31.475441 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:31 crc kubenswrapper[4678]: I1206 10:38:31.475442 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:31 crc kubenswrapper[4678]: E1206 10:38:31.476184 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:31 crc kubenswrapper[4678]: E1206 10:38:31.476398 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:32 crc kubenswrapper[4678]: I1206 10:38:32.475734 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:32 crc kubenswrapper[4678]: I1206 10:38:32.475799 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:32 crc kubenswrapper[4678]: E1206 10:38:32.475952 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:32 crc kubenswrapper[4678]: E1206 10:38:32.476157 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:33 crc kubenswrapper[4678]: I1206 10:38:33.475542 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:33 crc kubenswrapper[4678]: E1206 10:38:33.475695 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:33 crc kubenswrapper[4678]: I1206 10:38:33.475542 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:33 crc kubenswrapper[4678]: E1206 10:38:33.475922 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:34 crc kubenswrapper[4678]: I1206 10:38:34.475129 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:34 crc kubenswrapper[4678]: I1206 10:38:34.475208 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:34 crc kubenswrapper[4678]: E1206 10:38:34.475383 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:34 crc kubenswrapper[4678]: E1206 10:38:34.475572 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:35 crc kubenswrapper[4678]: E1206 10:38:35.464047 4678 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 06 10:38:35 crc kubenswrapper[4678]: I1206 10:38:35.475824 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:35 crc kubenswrapper[4678]: I1206 10:38:35.475840 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:35 crc kubenswrapper[4678]: E1206 10:38:35.477800 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:35 crc kubenswrapper[4678]: E1206 10:38:35.477858 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:35 crc kubenswrapper[4678]: E1206 10:38:35.572312 4678 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 10:38:36 crc kubenswrapper[4678]: I1206 10:38:36.475517 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:36 crc kubenswrapper[4678]: I1206 10:38:36.475556 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:36 crc kubenswrapper[4678]: E1206 10:38:36.475736 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:36 crc kubenswrapper[4678]: E1206 10:38:36.475852 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:37 crc kubenswrapper[4678]: I1206 10:38:37.475534 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:37 crc kubenswrapper[4678]: I1206 10:38:37.475637 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:37 crc kubenswrapper[4678]: E1206 10:38:37.475757 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:37 crc kubenswrapper[4678]: E1206 10:38:37.475906 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:38 crc kubenswrapper[4678]: I1206 10:38:38.475135 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:38 crc kubenswrapper[4678]: I1206 10:38:38.475195 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:38 crc kubenswrapper[4678]: E1206 10:38:38.475368 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:38 crc kubenswrapper[4678]: E1206 10:38:38.475514 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:38 crc kubenswrapper[4678]: I1206 10:38:38.698776 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:38:39 crc kubenswrapper[4678]: I1206 10:38:39.475642 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:39 crc kubenswrapper[4678]: I1206 10:38:39.475734 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:39 crc kubenswrapper[4678]: E1206 10:38:39.475886 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:39 crc kubenswrapper[4678]: E1206 10:38:39.475985 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:40 crc kubenswrapper[4678]: I1206 10:38:40.475424 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:40 crc kubenswrapper[4678]: I1206 10:38:40.475587 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:40 crc kubenswrapper[4678]: E1206 10:38:40.475628 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:40 crc kubenswrapper[4678]: E1206 10:38:40.475811 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:40 crc kubenswrapper[4678]: I1206 10:38:40.476395 4678 scope.go:117] "RemoveContainer" containerID="278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13" Dec 06 10:38:40 crc kubenswrapper[4678]: E1206 10:38:40.573688 4678 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 06 10:38:41 crc kubenswrapper[4678]: I1206 10:38:41.254450 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/1.log" Dec 06 10:38:41 crc kubenswrapper[4678]: I1206 10:38:41.254591 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerStarted","Data":"1f99ae1ba562ece348c05726186b8fa5e06897de886758153dfe02b1c5a8e84c"} Dec 06 10:38:41 crc kubenswrapper[4678]: I1206 10:38:41.288015 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-z5s25" podStartSLOduration=108.287988845 podStartE2EDuration="1m48.287988845s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:41.287939073 +0000 UTC m=+126.131370552" watchObservedRunningTime="2025-12-06 10:38:41.287988845 +0000 UTC m=+126.131420314" Dec 06 10:38:41 crc kubenswrapper[4678]: I1206 10:38:41.475975 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:41 crc kubenswrapper[4678]: I1206 10:38:41.476031 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:41 crc kubenswrapper[4678]: E1206 10:38:41.476195 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:41 crc kubenswrapper[4678]: E1206 10:38:41.476319 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:42 crc kubenswrapper[4678]: I1206 10:38:42.475099 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:42 crc kubenswrapper[4678]: I1206 10:38:42.475144 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:42 crc kubenswrapper[4678]: E1206 10:38:42.475364 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:42 crc kubenswrapper[4678]: E1206 10:38:42.475521 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:43 crc kubenswrapper[4678]: I1206 10:38:43.478681 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:43 crc kubenswrapper[4678]: I1206 10:38:43.478723 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:43 crc kubenswrapper[4678]: E1206 10:38:43.478894 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:43 crc kubenswrapper[4678]: E1206 10:38:43.479076 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:44 crc kubenswrapper[4678]: I1206 10:38:44.475756 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:44 crc kubenswrapper[4678]: E1206 10:38:44.475959 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 06 10:38:44 crc kubenswrapper[4678]: I1206 10:38:44.475930 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:44 crc kubenswrapper[4678]: E1206 10:38:44.476286 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 06 10:38:45 crc kubenswrapper[4678]: I1206 10:38:45.475105 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:45 crc kubenswrapper[4678]: E1206 10:38:45.475260 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 06 10:38:45 crc kubenswrapper[4678]: I1206 10:38:45.475554 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:45 crc kubenswrapper[4678]: E1206 10:38:45.477260 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-wfvj4" podUID="5fdea9d8-255a-4ec7-8095-4489828f4136" Dec 06 10:38:46 crc kubenswrapper[4678]: I1206 10:38:46.475334 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:38:46 crc kubenswrapper[4678]: I1206 10:38:46.475380 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:38:46 crc kubenswrapper[4678]: I1206 10:38:46.479277 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 10:38:46 crc kubenswrapper[4678]: I1206 10:38:46.479658 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 10:38:47 crc kubenswrapper[4678]: I1206 10:38:47.475698 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:38:47 crc kubenswrapper[4678]: I1206 10:38:47.476066 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:38:47 crc kubenswrapper[4678]: I1206 10:38:47.481331 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 10:38:47 crc kubenswrapper[4678]: I1206 10:38:47.481583 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 10:38:47 crc kubenswrapper[4678]: I1206 10:38:47.482709 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 10:38:47 crc kubenswrapper[4678]: I1206 10:38:47.482717 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 10:38:49 crc kubenswrapper[4678]: I1206 10:38:49.948787 4678 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.003996 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ck2d4"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.004862 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.008658 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rpv8w"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.009648 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.011072 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qbk7d"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.011812 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.013396 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.014086 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.032924 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.033816 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.034664 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.034984 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.035288 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.036412 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.036665 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.036724 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.036751 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.037461 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.037945 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.038252 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.038559 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.039788 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r5rng"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.040271 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5vk2z"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.040632 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.040747 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.064811 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.064826 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.064985 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.065445 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.065701 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.065987 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.066058 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.065993 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.066243 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.066279 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.066730 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.067094 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.067296 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.067539 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.067783 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.070105 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.075214 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.075868 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.076642 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.078335 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.078806 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.078956 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.081107 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.084450 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.085008 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.086314 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.086611 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.087601 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.089678 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.114798 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.116792 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.117634 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.118089 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.118321 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.118429 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ll9jl"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.118590 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.118756 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.118802 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.118902 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.119349 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.120029 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.120370 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.120431 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lff5h"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.120996 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.121634 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-4scxr"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.122051 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-4scxr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.122993 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.123813 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.125524 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.125660 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.125797 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126069 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126168 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7v92j"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126479 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126530 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126597 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126777 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126838 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rth98"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.126920 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.127161 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.127350 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.127726 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.127941 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7b5fs"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.128417 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.131649 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.133991 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.136986 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-node-pullsecrets\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137034 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-audit-dir\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137060 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137087 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk8x2\" (UniqueName: \"kubernetes.io/projected/4280a0ae-d3d2-463a-9876-86967c99d560-kube-api-access-pk8x2\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137110 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137129 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137153 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137176 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgkc8\" (UniqueName: \"kubernetes.io/projected/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-kube-api-access-xgkc8\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137196 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-serving-cert\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137218 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d500b1dd-9d43-49bf-923a-c671bb6543d3-auth-proxy-config\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137238 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbb66\" (UniqueName: \"kubernetes.io/projected/53361e7d-fe05-4f40-a442-307cb30ecd1c-kube-api-access-pbb66\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137258 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-encryption-config\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137277 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13bf0b82-57c8-46a0-99e2-cc6a975168b8-serving-cert\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137297 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-etcd-client\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137323 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137349 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/421d8420-5eec-4a2e-9bac-f508328fdb84-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-r85v9\" (UID: \"421d8420-5eec-4a2e-9bac-f508328fdb84\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137370 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxt65\" (UniqueName: \"kubernetes.io/projected/421d8420-5eec-4a2e-9bac-f508328fdb84-kube-api-access-vxt65\") pod \"cluster-samples-operator-665b6dd947-r85v9\" (UID: \"421d8420-5eec-4a2e-9bac-f508328fdb84\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137396 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137420 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-audit\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137441 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-config\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137472 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-client-ca\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137531 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0061f615-2fb9-4e17-8ba5-62d97e8070c8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137567 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137592 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-policies\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137613 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137636 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-config\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137656 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137674 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137694 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137727 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0061f615-2fb9-4e17-8ba5-62d97e8070c8-config\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137745 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkkv9\" (UniqueName: \"kubernetes.io/projected/0061f615-2fb9-4e17-8ba5-62d97e8070c8-kube-api-access-mkkv9\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137769 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137791 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137824 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w454n\" (UniqueName: \"kubernetes.io/projected/13bf0b82-57c8-46a0-99e2-cc6a975168b8-kube-api-access-w454n\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137862 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4m9r\" (UniqueName: \"kubernetes.io/projected/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-kube-api-access-s4m9r\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137882 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d500b1dd-9d43-49bf-923a-c671bb6543d3-config\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137901 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0061f615-2fb9-4e17-8ba5-62d97e8070c8-images\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137920 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-dir\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137942 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137963 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-serving-cert\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.137985 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-service-ca-bundle\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138009 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-client-ca\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138030 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-image-import-ca\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138055 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkbmg\" (UniqueName: \"kubernetes.io/projected/d500b1dd-9d43-49bf-923a-c671bb6543d3-kube-api-access-pkbmg\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138076 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d500b1dd-9d43-49bf-923a-c671bb6543d3-machine-approver-tls\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138098 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-etcd-serving-ca\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138220 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-config\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138253 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4280a0ae-d3d2-463a-9876-86967c99d560-serving-cert\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.138294 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-config\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.141673 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.142388 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-98427"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.142929 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.143190 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.154747 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.165163 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.185188 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.186014 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.186926 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.189183 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.192312 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.194436 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.194899 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.195770 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.195916 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.196105 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.196231 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.196356 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.197136 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.197270 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.197421 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.197614 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.197752 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.197924 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.198526 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.198629 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.198746 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.198940 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.199085 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.199172 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.199314 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.200653 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.201103 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.201316 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.201699 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.201974 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.202183 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.202438 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.213458 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-z4j5s"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.213969 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214182 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214297 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214409 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214556 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214651 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214701 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214770 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214844 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.214932 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.215053 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.222993 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.225512 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.225606 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.226007 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.226422 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.226528 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.227049 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.227079 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.227816 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.228541 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.228941 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.231106 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.231723 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.233862 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.234387 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.234630 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.235196 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.236060 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.236460 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.236546 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.237573 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.237743 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.238566 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.238634 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241671 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241694 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxt65\" (UniqueName: \"kubernetes.io/projected/421d8420-5eec-4a2e-9bac-f508328fdb84-kube-api-access-vxt65\") pod \"cluster-samples-operator-665b6dd947-r85v9\" (UID: \"421d8420-5eec-4a2e-9bac-f508328fdb84\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241718 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428bf8a-2963-4011-97e0-c9f55248194b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241736 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-audit\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241755 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b13d6d52-897c-4e08-ad3f-6150130d5f9f-images\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241773 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-service-ca-bundle\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241790 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl8zw\" (UniqueName: \"kubernetes.io/projected/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-kube-api-access-wl8zw\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241806 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-serving-cert\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241826 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-client-ca\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241843 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0061f615-2fb9-4e17-8ba5-62d97e8070c8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241859 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-config\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241877 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241921 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-policies\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241938 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241955 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-config\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241970 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.241987 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7428bf8a-2963-4011-97e0-c9f55248194b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242004 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-stats-auth\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242022 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242039 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242057 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0061f615-2fb9-4e17-8ba5-62d97e8070c8-config\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242077 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242103 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkkv9\" (UniqueName: \"kubernetes.io/projected/0061f615-2fb9-4e17-8ba5-62d97e8070c8-kube-api-access-mkkv9\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242122 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-metrics-certs\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242139 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242140 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87qs5\" (UniqueName: \"kubernetes.io/projected/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-kube-api-access-87qs5\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242396 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44xts\" (UniqueName: \"kubernetes.io/projected/f5b1144c-a578-4781-b54b-9971cc41fbe6-kube-api-access-44xts\") pod \"dns-operator-744455d44c-7v92j\" (UID: \"f5b1144c-a578-4781-b54b-9971cc41fbe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242434 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242456 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242473 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b13d6d52-897c-4e08-ad3f-6150130d5f9f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.242537 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w454n\" (UniqueName: \"kubernetes.io/projected/13bf0b82-57c8-46a0-99e2-cc6a975168b8-kube-api-access-w454n\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.243294 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-audit\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.243623 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.244047 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.244279 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245138 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-config\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245188 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-client-ca\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245449 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-dir\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245479 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4m9r\" (UniqueName: \"kubernetes.io/projected/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-kube-api-access-s4m9r\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245531 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d500b1dd-9d43-49bf-923a-c671bb6543d3-config\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245547 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0061f615-2fb9-4e17-8ba5-62d97e8070c8-images\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245562 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245600 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-serving-cert\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245616 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-service-ca-bundle\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245758 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkbmg\" (UniqueName: \"kubernetes.io/projected/d500b1dd-9d43-49bf-923a-c671bb6543d3-kube-api-access-pkbmg\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245782 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-client-ca\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245797 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-image-import-ca\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245859 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d500b1dd-9d43-49bf-923a-c671bb6543d3-machine-approver-tls\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245915 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.245985 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-etcd-serving-ca\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246032 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae8d3b71-e129-4eeb-b72c-617aa008538a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246051 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7428bf8a-2963-4011-97e0-c9f55248194b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246072 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b13d6d52-897c-4e08-ad3f-6150130d5f9f-proxy-tls\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246113 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4280a0ae-d3d2-463a-9876-86967c99d560-serving-cert\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246133 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ae8d3b71-e129-4eeb-b72c-617aa008538a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246155 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-default-certificate\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246238 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-config\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246258 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-config\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-node-pullsecrets\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246325 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-audit-dir\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246347 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xttwg\" (UniqueName: \"kubernetes.io/projected/ae8d3b71-e129-4eeb-b72c-617aa008538a-kube-api-access-xttwg\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246384 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246402 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk8x2\" (UniqueName: \"kubernetes.io/projected/4280a0ae-d3d2-463a-9876-86967c99d560-kube-api-access-pk8x2\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246420 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246452 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q84cd\" (UniqueName: \"kubernetes.io/projected/b13d6d52-897c-4e08-ad3f-6150130d5f9f-kube-api-access-q84cd\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246476 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgkc8\" (UniqueName: \"kubernetes.io/projected/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-kube-api-access-xgkc8\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246517 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246537 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246556 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-serving-cert\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246588 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d500b1dd-9d43-49bf-923a-c671bb6543d3-auth-proxy-config\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246605 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbb66\" (UniqueName: \"kubernetes.io/projected/53361e7d-fe05-4f40-a442-307cb30ecd1c-kube-api-access-pbb66\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246627 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f5b1144c-a578-4781-b54b-9971cc41fbe6-metrics-tls\") pod \"dns-operator-744455d44c-7v92j\" (UID: \"f5b1144c-a578-4781-b54b-9971cc41fbe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246645 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtvwc\" (UniqueName: \"kubernetes.io/projected/08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4-kube-api-access-dtvwc\") pod \"downloads-7954f5f757-4scxr\" (UID: \"08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4\") " pod="openshift-console/downloads-7954f5f757-4scxr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246678 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13bf0b82-57c8-46a0-99e2-cc6a975168b8-serving-cert\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246693 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-etcd-client\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246709 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-encryption-config\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246747 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246764 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/421d8420-5eec-4a2e-9bac-f508328fdb84-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-r85v9\" (UID: \"421d8420-5eec-4a2e-9bac-f508328fdb84\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.246779 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae8d3b71-e129-4eeb-b72c-617aa008538a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.247007 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-dir\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.247888 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d500b1dd-9d43-49bf-923a-c671bb6543d3-config\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.248703 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.248827 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0061f615-2fb9-4e17-8ba5-62d97e8070c8-images\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.248900 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-policies\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.249170 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.250156 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.250668 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.254850 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.257182 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.257797 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.258453 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.258692 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0061f615-2fb9-4e17-8ba5-62d97e8070c8-config\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.262183 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-service-ca-bundle\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.263689 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-client-ca\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.264543 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d500b1dd-9d43-49bf-923a-c671bb6543d3-auth-proxy-config\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.264728 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.266224 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13bf0b82-57c8-46a0-99e2-cc6a975168b8-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.266879 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-serving-cert\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.269967 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-config\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.270185 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.270212 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d500b1dd-9d43-49bf-923a-c671bb6543d3-machine-approver-tls\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.270466 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rrj9g"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.270826 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.271255 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wbcwh"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.271686 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-etcd-serving-ca\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.272050 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-image-import-ca\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.273505 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.273862 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.274053 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-node-pullsecrets\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.274682 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-audit-dir\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.275105 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-config\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.275267 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-config\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.275342 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.276269 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.277161 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-serving-cert\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.277445 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.280011 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.280532 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qn4db"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.280674 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.280918 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.282261 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.282652 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.282912 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.283154 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.283568 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.284058 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-etcd-client\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.284191 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4280a0ae-d3d2-463a-9876-86967c99d560-serving-cert\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.284293 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13bf0b82-57c8-46a0-99e2-cc6a975168b8-serving-cert\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.284421 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.284536 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-encryption-config\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.286799 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.296460 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.297743 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/0061f615-2fb9-4e17-8ba5-62d97e8070c8-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.298292 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/421d8420-5eec-4a2e-9bac-f508328fdb84-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-r85v9\" (UID: \"421d8420-5eec-4a2e-9bac-f508328fdb84\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.302226 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.302462 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.302534 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5vk2z"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.302948 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rpv8w"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.306196 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ck2d4"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.306950 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.307691 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.309187 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.310839 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qbk7d"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.312987 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.315594 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-4scxr"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.318436 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.322930 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-g8tzz"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.324059 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.324584 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rth98"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.325809 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.327259 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.334737 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lff5h"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.336855 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.338220 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7v92j"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.341138 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.342636 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.345510 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-m42c5"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.347044 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.347433 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428bf8a-2963-4011-97e0-c9f55248194b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.347467 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b13d6d52-897c-4e08-ad3f-6150130d5f9f-images\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.347502 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-service-ca-bundle\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.347753 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348026 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl8zw\" (UniqueName: \"kubernetes.io/projected/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-kube-api-access-wl8zw\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348070 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-serving-cert\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348100 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348239 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7428bf8a-2963-4011-97e0-c9f55248194b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348276 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-stats-auth\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348314 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-metrics-certs\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348348 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44xts\" (UniqueName: \"kubernetes.io/projected/f5b1144c-a578-4781-b54b-9971cc41fbe6-kube-api-access-44xts\") pod \"dns-operator-744455d44c-7v92j\" (UID: \"f5b1144c-a578-4781-b54b-9971cc41fbe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348368 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87qs5\" (UniqueName: \"kubernetes.io/projected/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-kube-api-access-87qs5\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.348804 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b13d6d52-897c-4e08-ad3f-6150130d5f9f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349098 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae8d3b71-e129-4eeb-b72c-617aa008538a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349131 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7428bf8a-2963-4011-97e0-c9f55248194b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349150 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b13d6d52-897c-4e08-ad3f-6150130d5f9f-proxy-tls\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349185 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ae8d3b71-e129-4eeb-b72c-617aa008538a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349210 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-default-certificate\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349246 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xttwg\" (UniqueName: \"kubernetes.io/projected/ae8d3b71-e129-4eeb-b72c-617aa008538a-kube-api-access-xttwg\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349279 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q84cd\" (UniqueName: \"kubernetes.io/projected/b13d6d52-897c-4e08-ad3f-6150130d5f9f-kube-api-access-q84cd\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349332 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f5b1144c-a578-4781-b54b-9971cc41fbe6-metrics-tls\") pod \"dns-operator-744455d44c-7v92j\" (UID: \"f5b1144c-a578-4781-b54b-9971cc41fbe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349624 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtvwc\" (UniqueName: \"kubernetes.io/projected/08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4-kube-api-access-dtvwc\") pod \"downloads-7954f5f757-4scxr\" (UID: \"08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4\") " pod="openshift-console/downloads-7954f5f757-4scxr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae8d3b71-e129-4eeb-b72c-617aa008538a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349847 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.349974 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b13d6d52-897c-4e08-ad3f-6150130d5f9f-auth-proxy-config\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.350840 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7428bf8a-2963-4011-97e0-c9f55248194b-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.351285 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-serving-cert\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.352588 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ae8d3b71-e129-4eeb-b72c-617aa008538a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.353103 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ae8d3b71-e129-4eeb-b72c-617aa008538a-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.353623 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.354040 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f5b1144c-a578-4781-b54b-9971cc41fbe6-metrics-tls\") pod \"dns-operator-744455d44c-7v92j\" (UID: \"f5b1144c-a578-4781-b54b-9971cc41fbe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.355192 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.356370 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r5rng"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.357472 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7b5fs"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.358655 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ll9jl"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.364899 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-m42c5"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.367833 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.368831 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.370640 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.372056 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.373250 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-98427"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.374256 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.375348 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.376677 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.377673 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.378660 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.379664 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.380884 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rtcvz"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.381654 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rrj9g"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.381679 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.382681 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.383730 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qn4db"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.384752 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2wz6s"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.386081 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.388011 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.388224 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wbcwh"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.389003 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rtcvz"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.390383 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2wz6s"] Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.407563 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.414120 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7428bf8a-2963-4011-97e0-c9f55248194b-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.435819 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.448988 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.469879 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.487507 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.507906 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.528325 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.547254 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.568354 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.586619 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.606454 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.627401 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.634162 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-default-certificate\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.647864 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.668437 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.688197 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.694358 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-stats-auth\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.708707 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.723467 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-metrics-certs\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.727462 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.728251 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/b13d6d52-897c-4e08-ad3f-6150130d5f9f-images\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.748218 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.760122 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-service-ca-bundle\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.768412 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.788171 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.795412 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b13d6d52-897c-4e08-ad3f-6150130d5f9f-proxy-tls\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.808180 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.867168 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.887504 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.907850 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.928218 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.947195 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.966971 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 10:38:50 crc kubenswrapper[4678]: I1206 10:38:50.987527 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.008887 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.028836 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.047968 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.067803 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.087438 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.108766 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.127229 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.149334 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.168087 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.188324 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.228732 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.236440 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxt65\" (UniqueName: \"kubernetes.io/projected/421d8420-5eec-4a2e-9bac-f508328fdb84-kube-api-access-vxt65\") pod \"cluster-samples-operator-665b6dd947-r85v9\" (UID: \"421d8420-5eec-4a2e-9bac-f508328fdb84\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.246084 4678 request.go:700] Waited for 1.001796272s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serviceaccount-dockercfg-rq7zk&limit=500&resourceVersion=0 Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.248684 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.268148 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.287179 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.308241 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.347744 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.350178 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w454n\" (UniqueName: \"kubernetes.io/projected/13bf0b82-57c8-46a0-99e2-cc6a975168b8-kube-api-access-w454n\") pod \"authentication-operator-69f744f599-rpv8w\" (UID: \"13bf0b82-57c8-46a0-99e2-cc6a975168b8\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.368836 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.388848 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.425989 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4m9r\" (UniqueName: \"kubernetes.io/projected/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-kube-api-access-s4m9r\") pod \"controller-manager-879f6c89f-qbk7d\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.428585 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.436015 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.448646 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.490716 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkkv9\" (UniqueName: \"kubernetes.io/projected/0061f615-2fb9-4e17-8ba5-62d97e8070c8-kube-api-access-mkkv9\") pod \"machine-api-operator-5694c8668f-r5rng\" (UID: \"0061f615-2fb9-4e17-8ba5-62d97e8070c8\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.511956 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkbmg\" (UniqueName: \"kubernetes.io/projected/d500b1dd-9d43-49bf-923a-c671bb6543d3-kube-api-access-pkbmg\") pod \"machine-approver-56656f9798-h8wxt\" (UID: \"d500b1dd-9d43-49bf-923a-c671bb6543d3\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.534294 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk8x2\" (UniqueName: \"kubernetes.io/projected/4280a0ae-d3d2-463a-9876-86967c99d560-kube-api-access-pk8x2\") pod \"route-controller-manager-6576b87f9c-64p9r\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.546061 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgkc8\" (UniqueName: \"kubernetes.io/projected/e1cd25ff-72f0-414f-a4e6-51c0d5eb5662-kube-api-access-xgkc8\") pod \"apiserver-76f77b778f-ck2d4\" (UID: \"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662\") " pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.559751 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.566254 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbb66\" (UniqueName: \"kubernetes.io/projected/53361e7d-fe05-4f40-a442-307cb30ecd1c-kube-api-access-pbb66\") pod \"oauth-openshift-558db77b4-5vk2z\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.568903 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.569951 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.592226 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.607799 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.616124 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.638125 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.638368 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.647208 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.648768 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9"] Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.670060 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.690028 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.696416 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.708158 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.715657 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.727901 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.747903 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.767980 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.789873 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.798987 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rpv8w"] Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.807814 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.827736 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.837763 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:38:51 crc kubenswrapper[4678]: W1206 10:38:51.839993 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13bf0b82_57c8_46a0_99e2_cc6a975168b8.slice/crio-8cca14fb27080a27c04ce16512ee72c4a0ad754dfed5273a31970ce43ab6858e WatchSource:0}: Error finding container 8cca14fb27080a27c04ce16512ee72c4a0ad754dfed5273a31970ce43ab6858e: Status 404 returned error can't find the container with id 8cca14fb27080a27c04ce16512ee72c4a0ad754dfed5273a31970ce43ab6858e Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.847774 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.867359 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.877771 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qbk7d"] Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.887661 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.910312 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.922874 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r"] Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.927840 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.951538 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.967338 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.971522 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-r5rng"] Dec 06 10:38:51 crc kubenswrapper[4678]: I1206 10:38:51.989625 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.006437 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5vk2z"] Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.006786 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.027233 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.046706 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.066792 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.124543 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44xts\" (UniqueName: \"kubernetes.io/projected/f5b1144c-a578-4781-b54b-9971cc41fbe6-kube-api-access-44xts\") pod \"dns-operator-744455d44c-7v92j\" (UID: \"f5b1144c-a578-4781-b54b-9971cc41fbe6\") " pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.143260 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7428bf8a-2963-4011-97e0-c9f55248194b-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-f6m8w\" (UID: \"7428bf8a-2963-4011-97e0-c9f55248194b\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.163797 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.164793 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87qs5\" (UniqueName: \"kubernetes.io/projected/6508c5a9-1514-4fe7-88cf-9d5f03af9ff9-kube-api-access-87qs5\") pod \"openshift-config-operator-7777fb866f-rth98\" (UID: \"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.182634 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ae8d3b71-e129-4eeb-b72c-617aa008538a-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.184615 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.201954 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xttwg\" (UniqueName: \"kubernetes.io/projected/ae8d3b71-e129-4eeb-b72c-617aa008538a-kube-api-access-xttwg\") pod \"cluster-image-registry-operator-dc59b4c8b-lfrsj\" (UID: \"ae8d3b71-e129-4eeb-b72c-617aa008538a\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.234445 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q84cd\" (UniqueName: \"kubernetes.io/projected/b13d6d52-897c-4e08-ad3f-6150130d5f9f-kube-api-access-q84cd\") pod \"machine-config-operator-74547568cd-p79pr\" (UID: \"b13d6d52-897c-4e08-ad3f-6150130d5f9f\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.246213 4678 request.go:700] Waited for 1.864241982s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/secrets?fieldSelector=metadata.name%3Dcanary-serving-cert&limit=500&resourceVersion=0 Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.247712 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.252859 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtvwc\" (UniqueName: \"kubernetes.io/projected/08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4-kube-api-access-dtvwc\") pod \"downloads-7954f5f757-4scxr\" (UID: \"08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4\") " pod="openshift-console/downloads-7954f5f757-4scxr" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.267393 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.288128 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.308027 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.313907 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" event={"ID":"d500b1dd-9d43-49bf-923a-c671bb6543d3","Type":"ContainerStarted","Data":"fa284c7c7033a28058f6013ce1e2709180850c068772503a2881b77e6165a61c"} Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.314991 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" event={"ID":"13bf0b82-57c8-46a0-99e2-cc6a975168b8","Type":"ContainerStarted","Data":"8cca14fb27080a27c04ce16512ee72c4a0ad754dfed5273a31970ce43ab6858e"} Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.327470 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.347878 4678 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 10:38:52 crc kubenswrapper[4678]: I1206 10:38:52.367954 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.367279 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-4scxr" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.370879 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.371980 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.376151 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.385064 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl8zw\" (UniqueName: \"kubernetes.io/projected/40b9fa22-f2c6-4fb6-993a-65b61c12fed7-kube-api-access-wl8zw\") pod \"router-default-5444994796-z4j5s\" (UID: \"40b9fa22-f2c6-4fb6-993a-65b61c12fed7\") " pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.393508 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.396141 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/140fee33-c778-4f79-941e-430f5ae44bd9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.396170 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-trusted-ca\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.396861 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-bound-sa-token\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.396889 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-registry-certificates\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.400653 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.400794 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/140fee33-c778-4f79-941e-430f5ae44bd9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.400847 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-registry-tls\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.400872 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcgcz\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-kube-api-access-hcgcz\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.401177 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:53.901161423 +0000 UTC m=+138.744592862 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.406446 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" event={"ID":"0061f615-2fb9-4e17-8ba5-62d97e8070c8","Type":"ContainerStarted","Data":"85449d524a9b9387c5036fb58e31827b5c4b11691a76cdc3678abff648902429"} Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.410618 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" event={"ID":"53361e7d-fe05-4f40-a442-307cb30ecd1c","Type":"ContainerStarted","Data":"f721f994eb94a9212768c6dfff627e48e4b0631edadbab6307a44d1803963e32"} Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.411452 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ck2d4"] Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.413599 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" event={"ID":"4280a0ae-d3d2-463a-9876-86967c99d560","Type":"ContainerStarted","Data":"5b27a6e629c7be4c6e311dac7e221374117bf6ceb8b9b5433b338f81bf1b09a2"} Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.415189 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" event={"ID":"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732","Type":"ContainerStarted","Data":"1dd225ef85e8fda491ee1f4e768e96a5aee9a05cf28cb87c773520a78a1b82e0"} Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.504029 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.505718 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.005674311 +0000 UTC m=+138.849105750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.514082 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b0941dc-500a-4284-9cc4-66b250feee93-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvssp\" (UID: \"9b0941dc-500a-4284-9cc4-66b250feee93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.514242 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-serving-cert\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.514770 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kssc7\" (UniqueName: \"kubernetes.io/projected/ed43fec1-dba7-4db3-bb66-f161c4c9c622-kube-api-access-kssc7\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.514936 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvhww\" (UniqueName: \"kubernetes.io/projected/f13f288c-c8dc-4df8-b89b-c325d088b30b-kube-api-access-lvhww\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.515096 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d00e05-6326-4f98-add6-ca019b7791f4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.515209 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460d0786-6a9e-4d3e-ae41-2cb110d42336-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.515751 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.515871 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wng74\" (UniqueName: \"kubernetes.io/projected/72675578-a2c8-4f7b-a36b-99c5996703b5-kube-api-access-wng74\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.515996 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4777f215-78ef-47d4-be38-0e2a750168fe-config-volume\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.516156 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-config\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.516352 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-serving-cert\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.516915 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb-cert\") pod \"ingress-canary-rtcvz\" (UID: \"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb\") " pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.517069 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-oauth-serving-cert\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.517368 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkbpf\" (UniqueName: \"kubernetes.io/projected/73a80455-9c8e-4061-bf7b-6afb42555cd6-kube-api-access-lkbpf\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.517996 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqgxs\" (UniqueName: \"kubernetes.io/projected/b49b0624-3162-4a68-b636-cd57545d7016-kube-api-access-cqgxs\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518067 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-encryption-config\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518085 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4777f215-78ef-47d4-be38-0e2a750168fe-secret-volume\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518102 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed43fec1-dba7-4db3-bb66-f161c4c9c622-config-volume\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518120 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0bbdcc6e-f0b1-4541-877c-9bec206a2897-tmpfs\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518138 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss96x\" (UniqueName: \"kubernetes.io/projected/0bbdcc6e-f0b1-4541-877c-9bec206a2897-kube-api-access-ss96x\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518156 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/47911926-004f-468c-8d77-e1acefac68c3-certs\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518183 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-config\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518204 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518226 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl9m4\" (UniqueName: \"kubernetes.io/projected/9b0941dc-500a-4284-9cc4-66b250feee93-kube-api-access-gl9m4\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvssp\" (UID: \"9b0941dc-500a-4284-9cc4-66b250feee93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518245 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8wc7\" (UniqueName: \"kubernetes.io/projected/93940ef4-2998-4942-9c46-68e92af94436-kube-api-access-h8wc7\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518299 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-oauth-config\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518317 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a6d7574-8808-4c7d-aef1-9df452c83f27-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518379 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-service-ca\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518398 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518419 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36d00e05-6326-4f98-add6-ca019b7791f4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518437 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed43fec1-dba7-4db3-bb66-f161c4c9c622-metrics-tls\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518456 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzblk\" (UniqueName: \"kubernetes.io/projected/085d0a79-150c-45f0-993a-94c8b9d1811f-kube-api-access-jzblk\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518473 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518501 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73a80455-9c8e-4061-bf7b-6afb42555cd6-metrics-tls\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518517 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0bbdcc6e-f0b1-4541-877c-9bec206a2897-apiservice-cert\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518558 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518576 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-registration-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518596 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhn42\" (UniqueName: \"kubernetes.io/projected/be306aba-f697-4dfb-b87b-7beb769f6654-kube-api-access-lhn42\") pod \"multus-admission-controller-857f4d67dd-qn4db\" (UID: \"be306aba-f697-4dfb-b87b-7beb769f6654\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518616 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-serving-cert\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518634 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73a80455-9c8e-4061-bf7b-6afb42555cd6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518651 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-mountpoint-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518672 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a80455-9c8e-4061-bf7b-6afb42555cd6-trusted-ca\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518691 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdkpg\" (UniqueName: \"kubernetes.io/projected/1438bcb1-f3fb-4979-91ba-2bce2952edb9-kube-api-access-pdkpg\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518707 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3439565e-e37c-4962-a269-ddb8a48b3f43-audit-dir\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518725 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4kgq\" (UniqueName: \"kubernetes.io/projected/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-kube-api-access-w4kgq\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518776 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-registry-tls\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518795 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcgcz\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-kube-api-access-hcgcz\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518812 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-etcd-client\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518830 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc86x\" (UniqueName: \"kubernetes.io/projected/47911926-004f-468c-8d77-e1acefac68c3-kube-api-access-dc86x\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518847 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518885 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438bcb1-f3fb-4979-91ba-2bce2952edb9-config\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518923 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518946 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwwbs\" (UniqueName: \"kubernetes.io/projected/a0c46190-8e1a-4068-ba9f-9d3751787308-kube-api-access-bwwbs\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518970 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5e602f2b-acc1-433f-852d-62aad67403f3-srv-cert\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.518991 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-csi-data-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519010 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-plugins-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519030 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b49b0624-3162-4a68-b636-cd57545d7016-srv-cert\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519049 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-client\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519075 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-ca\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519093 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/140fee33-c778-4f79-941e-430f5ae44bd9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519111 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-trusted-ca\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519127 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dpmn\" (UniqueName: \"kubernetes.io/projected/4777f215-78ef-47d4-be38-0e2a750168fe-kube-api-access-5dpmn\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519159 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a6d7574-8808-4c7d-aef1-9df452c83f27-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519177 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmhdn\" (UniqueName: \"kubernetes.io/projected/9a6d7574-8808-4c7d-aef1-9df452c83f27-kube-api-access-xmhdn\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519207 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/be306aba-f697-4dfb-b87b-7beb769f6654-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qn4db\" (UID: \"be306aba-f697-4dfb-b87b-7beb769f6654\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519224 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9063a39-025a-4fac-a196-eed330238262-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pl9h6\" (UID: \"f9063a39-025a-4fac-a196-eed330238262\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519240 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0bbdcc6e-f0b1-4541-877c-9bec206a2897-webhook-cert\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519260 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-bound-sa-token\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519276 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-config\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519292 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8xsx\" (UniqueName: \"kubernetes.io/projected/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-kube-api-access-m8xsx\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519310 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93940ef4-2998-4942-9c46-68e92af94436-config\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519337 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-proxy-tls\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519367 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-registry-certificates\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519407 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1438bcb1-f3fb-4979-91ba-2bce2952edb9-trusted-ca\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519428 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-895zv\" (UniqueName: \"kubernetes.io/projected/6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb-kube-api-access-895zv\") pod \"ingress-canary-rtcvz\" (UID: \"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb\") " pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519446 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519472 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-socket-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519507 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460d0786-6a9e-4d3e-ae41-2cb110d42336-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519545 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519561 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1438bcb1-f3fb-4979-91ba-2bce2952edb9-serving-cert\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519578 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5e602f2b-acc1-433f-852d-62aad67403f3-profile-collector-cert\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519597 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckthp\" (UniqueName: \"kubernetes.io/projected/5e602f2b-acc1-433f-852d-62aad67403f3-kube-api-access-ckthp\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519614 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d00e05-6326-4f98-add6-ca019b7791f4-config\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519644 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-248c9\" (UniqueName: \"kubernetes.io/projected/f9063a39-025a-4fac-a196-eed330238262-kube-api-access-248c9\") pod \"package-server-manager-789f6589d5-pl9h6\" (UID: \"f9063a39-025a-4fac-a196-eed330238262\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519673 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-trusted-ca-bundle\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519693 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qx7\" (UniqueName: \"kubernetes.io/projected/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-kube-api-access-l9qx7\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519714 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds77q\" (UniqueName: \"kubernetes.io/projected/460d0786-6a9e-4d3e-ae41-2cb110d42336-kube-api-access-ds77q\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519734 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-audit-policies\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519751 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b49b0624-3162-4a68-b636-cd57545d7016-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519780 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxxkr\" (UniqueName: \"kubernetes.io/projected/3439565e-e37c-4962-a269-ddb8a48b3f43-kube-api-access-fxxkr\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519799 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdsq8\" (UniqueName: \"kubernetes.io/projected/2cf70a61-94e0-4b41-9823-f43bb83f2dcf-kube-api-access-cdsq8\") pod \"migrator-59844c95c7-hfsx2\" (UID: \"2cf70a61-94e0-4b41-9823-f43bb83f2dcf\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519815 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/085d0a79-150c-45f0-993a-94c8b9d1811f-signing-cabundle\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519836 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519867 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/140fee33-c778-4f79-941e-430f5ae44bd9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519884 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-service-ca\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519899 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/085d0a79-150c-45f0-993a-94c8b9d1811f-signing-key\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519916 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/47911926-004f-468c-8d77-e1acefac68c3-node-bootstrap-token\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.519933 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93940ef4-2998-4942-9c46-68e92af94436-serving-cert\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.522879 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.022865308 +0000 UTC m=+138.866296747 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.528840 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/140fee33-c778-4f79-941e-430f5ae44bd9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.531153 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-registry-tls\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.534402 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/140fee33-c778-4f79-941e-430f5ae44bd9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.535473 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-registry-certificates\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.547523 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcgcz\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-kube-api-access-hcgcz\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.550667 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-bound-sa-token\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.554205 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-trusted-ca\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.621576 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.621796 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438bcb1-f3fb-4979-91ba-2bce2952edb9-config\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.621834 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.121799606 +0000 UTC m=+138.965231045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623001 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1438bcb1-f3fb-4979-91ba-2bce2952edb9-config\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623060 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623121 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwwbs\" (UniqueName: \"kubernetes.io/projected/a0c46190-8e1a-4068-ba9f-9d3751787308-kube-api-access-bwwbs\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623143 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5e602f2b-acc1-433f-852d-62aad67403f3-srv-cert\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623169 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-csi-data-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623187 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b49b0624-3162-4a68-b636-cd57545d7016-srv-cert\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623211 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-plugins-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623230 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-client\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623252 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-ca\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623272 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dpmn\" (UniqueName: \"kubernetes.io/projected/4777f215-78ef-47d4-be38-0e2a750168fe-kube-api-access-5dpmn\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623288 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a6d7574-8808-4c7d-aef1-9df452c83f27-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmhdn\" (UniqueName: \"kubernetes.io/projected/9a6d7574-8808-4c7d-aef1-9df452c83f27-kube-api-access-xmhdn\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623336 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9063a39-025a-4fac-a196-eed330238262-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pl9h6\" (UID: \"f9063a39-025a-4fac-a196-eed330238262\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623357 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0bbdcc6e-f0b1-4541-877c-9bec206a2897-webhook-cert\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623375 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/be306aba-f697-4dfb-b87b-7beb769f6654-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qn4db\" (UID: \"be306aba-f697-4dfb-b87b-7beb769f6654\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623397 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-config\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623416 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8xsx\" (UniqueName: \"kubernetes.io/projected/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-kube-api-access-m8xsx\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623436 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93940ef4-2998-4942-9c46-68e92af94436-config\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623455 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-proxy-tls\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623499 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1438bcb1-f3fb-4979-91ba-2bce2952edb9-trusted-ca\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623523 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-895zv\" (UniqueName: \"kubernetes.io/projected/6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb-kube-api-access-895zv\") pod \"ingress-canary-rtcvz\" (UID: \"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb\") " pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623547 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623567 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-socket-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623596 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460d0786-6a9e-4d3e-ae41-2cb110d42336-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623624 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623641 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1438bcb1-f3fb-4979-91ba-2bce2952edb9-serving-cert\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623656 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5e602f2b-acc1-433f-852d-62aad67403f3-profile-collector-cert\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckthp\" (UniqueName: \"kubernetes.io/projected/5e602f2b-acc1-433f-852d-62aad67403f3-kube-api-access-ckthp\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623701 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d00e05-6326-4f98-add6-ca019b7791f4-config\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623722 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-248c9\" (UniqueName: \"kubernetes.io/projected/f9063a39-025a-4fac-a196-eed330238262-kube-api-access-248c9\") pod \"package-server-manager-789f6589d5-pl9h6\" (UID: \"f9063a39-025a-4fac-a196-eed330238262\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623746 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-trusted-ca-bundle\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623763 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qx7\" (UniqueName: \"kubernetes.io/projected/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-kube-api-access-l9qx7\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623783 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds77q\" (UniqueName: \"kubernetes.io/projected/460d0786-6a9e-4d3e-ae41-2cb110d42336-kube-api-access-ds77q\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623801 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-audit-policies\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623819 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b49b0624-3162-4a68-b636-cd57545d7016-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623853 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxxkr\" (UniqueName: \"kubernetes.io/projected/3439565e-e37c-4962-a269-ddb8a48b3f43-kube-api-access-fxxkr\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdsq8\" (UniqueName: \"kubernetes.io/projected/2cf70a61-94e0-4b41-9823-f43bb83f2dcf-kube-api-access-cdsq8\") pod \"migrator-59844c95c7-hfsx2\" (UID: \"2cf70a61-94e0-4b41-9823-f43bb83f2dcf\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623890 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/085d0a79-150c-45f0-993a-94c8b9d1811f-signing-cabundle\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623914 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623935 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-service-ca\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623953 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/085d0a79-150c-45f0-993a-94c8b9d1811f-signing-key\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623974 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/47911926-004f-468c-8d77-e1acefac68c3-node-bootstrap-token\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.623989 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93940ef4-2998-4942-9c46-68e92af94436-serving-cert\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.624248 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b0941dc-500a-4284-9cc4-66b250feee93-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvssp\" (UID: \"9b0941dc-500a-4284-9cc4-66b250feee93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.624282 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-serving-cert\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.624304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kssc7\" (UniqueName: \"kubernetes.io/projected/ed43fec1-dba7-4db3-bb66-f161c4c9c622-kube-api-access-kssc7\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.624345 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvhww\" (UniqueName: \"kubernetes.io/projected/f13f288c-c8dc-4df8-b89b-c325d088b30b-kube-api-access-lvhww\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.624417 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d00e05-6326-4f98-add6-ca019b7791f4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.624440 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460d0786-6a9e-4d3e-ae41-2cb110d42336-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.624456 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.625548 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wng74\" (UniqueName: \"kubernetes.io/projected/72675578-a2c8-4f7b-a36b-99c5996703b5-kube-api-access-wng74\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.625598 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4777f215-78ef-47d4-be38-0e2a750168fe-config-volume\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.625617 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-config\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.629027 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-serving-cert\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.629050 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb-cert\") pod \"ingress-canary-rtcvz\" (UID: \"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb\") " pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.629078 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-oauth-serving-cert\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.629164 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkbpf\" (UniqueName: \"kubernetes.io/projected/73a80455-9c8e-4061-bf7b-6afb42555cd6-kube-api-access-lkbpf\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.629466 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqgxs\" (UniqueName: \"kubernetes.io/projected/b49b0624-3162-4a68-b636-cd57545d7016-kube-api-access-cqgxs\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.629870 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5e602f2b-acc1-433f-852d-62aad67403f3-srv-cert\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631151 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-encryption-config\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631168 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9063a39-025a-4fac-a196-eed330238262-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pl9h6\" (UID: \"f9063a39-025a-4fac-a196-eed330238262\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631182 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4777f215-78ef-47d4-be38-0e2a750168fe-secret-volume\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631268 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed43fec1-dba7-4db3-bb66-f161c4c9c622-config-volume\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631296 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0bbdcc6e-f0b1-4541-877c-9bec206a2897-tmpfs\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631332 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss96x\" (UniqueName: \"kubernetes.io/projected/0bbdcc6e-f0b1-4541-877c-9bec206a2897-kube-api-access-ss96x\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631366 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/47911926-004f-468c-8d77-e1acefac68c3-certs\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631393 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-config\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631426 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631453 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl9m4\" (UniqueName: \"kubernetes.io/projected/9b0941dc-500a-4284-9cc4-66b250feee93-kube-api-access-gl9m4\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvssp\" (UID: \"9b0941dc-500a-4284-9cc4-66b250feee93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631478 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8wc7\" (UniqueName: \"kubernetes.io/projected/93940ef4-2998-4942-9c46-68e92af94436-kube-api-access-h8wc7\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631528 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-oauth-config\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631551 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a6d7574-8808-4c7d-aef1-9df452c83f27-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631577 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-service-ca\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631601 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631624 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36d00e05-6326-4f98-add6-ca019b7791f4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631654 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed43fec1-dba7-4db3-bb66-f161c4c9c622-metrics-tls\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631676 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzblk\" (UniqueName: \"kubernetes.io/projected/085d0a79-150c-45f0-993a-94c8b9d1811f-kube-api-access-jzblk\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631702 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631741 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73a80455-9c8e-4061-bf7b-6afb42555cd6-metrics-tls\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631786 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0bbdcc6e-f0b1-4541-877c-9bec206a2897-apiservice-cert\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631916 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-registration-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.631964 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhn42\" (UniqueName: \"kubernetes.io/projected/be306aba-f697-4dfb-b87b-7beb769f6654-kube-api-access-lhn42\") pod \"multus-admission-controller-857f4d67dd-qn4db\" (UID: \"be306aba-f697-4dfb-b87b-7beb769f6654\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632005 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632059 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-serving-cert\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632098 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73a80455-9c8e-4061-bf7b-6afb42555cd6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632159 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-mountpoint-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632233 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a80455-9c8e-4061-bf7b-6afb42555cd6-trusted-ca\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdkpg\" (UniqueName: \"kubernetes.io/projected/1438bcb1-f3fb-4979-91ba-2bce2952edb9-kube-api-access-pdkpg\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632286 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3439565e-e37c-4962-a269-ddb8a48b3f43-audit-dir\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632309 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4kgq\" (UniqueName: \"kubernetes.io/projected/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-kube-api-access-w4kgq\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632344 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-etcd-client\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632367 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc86x\" (UniqueName: \"kubernetes.io/projected/47911926-004f-468c-8d77-e1acefac68c3-kube-api-access-dc86x\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.632390 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.633298 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.633843 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed43fec1-dba7-4db3-bb66-f161c4c9c622-config-volume\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.634136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0bbdcc6e-f0b1-4541-877c-9bec206a2897-tmpfs\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.636637 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.637751 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.637981 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4777f215-78ef-47d4-be38-0e2a750168fe-secret-volume\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.640659 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-client\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.640734 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-csi-data-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.642705 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-plugins-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.644764 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-service-ca\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.645555 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.646266 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/47911926-004f-468c-8d77-e1acefac68c3-certs\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.646830 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-config\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.647279 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-socket-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.648055 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b49b0624-3162-4a68-b636-cd57545d7016-srv-cert\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.648347 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36d00e05-6326-4f98-add6-ca019b7791f4-config\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.651321 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.667188 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.167167118 +0000 UTC m=+139.010598547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.670701 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-ca\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.672101 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-trusted-ca-bundle\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.672653 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-audit-policies\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.672854 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/085d0a79-150c-45f0-993a-94c8b9d1811f-signing-cabundle\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.681477 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-etcd-service-ca\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.687943 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93940ef4-2998-4942-9c46-68e92af94436-config\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.693993 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.695997 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9a6d7574-8808-4c7d-aef1-9df452c83f27-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.697399 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-mountpoint-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.698453 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/73a80455-9c8e-4061-bf7b-6afb42555cd6-trusted-ca\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.703287 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3439565e-e37c-4962-a269-ddb8a48b3f43-audit-dir\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.706720 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1438bcb1-f3fb-4979-91ba-2bce2952edb9-trusted-ca\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.720585 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36d00e05-6326-4f98-add6-ca019b7791f4-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.720702 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-oauth-config\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.725959 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss96x\" (UniqueName: \"kubernetes.io/projected/0bbdcc6e-f0b1-4541-877c-9bec206a2897-kube-api-access-ss96x\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.728368 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed43fec1-dba7-4db3-bb66-f161c4c9c622-metrics-tls\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.728831 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b49b0624-3162-4a68-b636-cd57545d7016-profile-collector-cert\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.728829 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-config\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.729210 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0bbdcc6e-f0b1-4541-877c-9bec206a2897-webhook-cert\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.731530 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/a0c46190-8e1a-4068-ba9f-9d3751787308-registration-dir\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.731774 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.733417 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.733531 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a6d7574-8808-4c7d-aef1-9df452c83f27-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.735631 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-serving-cert\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.736031 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-etcd-client\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.736740 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-proxy-tls\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.741293 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460d0786-6a9e-4d3e-ae41-2cb110d42336-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.742084 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-oauth-serving-cert\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.742604 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460d0786-6a9e-4d3e-ae41-2cb110d42336-config\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.742767 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-config\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.743319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3439565e-e37c-4962-a269-ddb8a48b3f43-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.745104 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.245066898 +0000 UTC m=+139.088498337 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.745340 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzblk\" (UniqueName: \"kubernetes.io/projected/085d0a79-150c-45f0-993a-94c8b9d1811f-kube-api-access-jzblk\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.746042 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdsq8\" (UniqueName: \"kubernetes.io/projected/2cf70a61-94e0-4b41-9823-f43bb83f2dcf-kube-api-access-cdsq8\") pod \"migrator-59844c95c7-hfsx2\" (UID: \"2cf70a61-94e0-4b41-9823-f43bb83f2dcf\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.748200 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4777f215-78ef-47d4-be38-0e2a750168fe-config-volume\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.749664 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b0941dc-500a-4284-9cc4-66b250feee93-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvssp\" (UID: \"9b0941dc-500a-4284-9cc4-66b250feee93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.749919 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/73a80455-9c8e-4061-bf7b-6afb42555cd6-metrics-tls\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.750142 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwwbs\" (UniqueName: \"kubernetes.io/projected/a0c46190-8e1a-4068-ba9f-9d3751787308-kube-api-access-bwwbs\") pod \"csi-hostpathplugin-2wz6s\" (UID: \"a0c46190-8e1a-4068-ba9f-9d3751787308\") " pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.750367 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxxkr\" (UniqueName: \"kubernetes.io/projected/3439565e-e37c-4962-a269-ddb8a48b3f43-kube-api-access-fxxkr\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.750758 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl9m4\" (UniqueName: \"kubernetes.io/projected/9b0941dc-500a-4284-9cc4-66b250feee93-kube-api-access-gl9m4\") pod \"control-plane-machine-set-operator-78cbb6b69f-dvssp\" (UID: \"9b0941dc-500a-4284-9cc4-66b250feee93\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.751066 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-encryption-config\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.751173 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/085d0a79-150c-45f0-993a-94c8b9d1811f-signing-key\") pod \"service-ca-9c57cc56f-rrj9g\" (UID: \"085d0a79-150c-45f0-993a-94c8b9d1811f\") " pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.751699 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-serving-cert\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.752191 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/93940ef4-2998-4942-9c46-68e92af94436-serving-cert\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.752241 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1438bcb1-f3fb-4979-91ba-2bce2952edb9-serving-cert\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.752743 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0bbdcc6e-f0b1-4541-877c-9bec206a2897-apiservice-cert\") pod \"packageserver-d55dfcdfc-ksqhs\" (UID: \"0bbdcc6e-f0b1-4541-877c-9bec206a2897\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.753293 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8wc7\" (UniqueName: \"kubernetes.io/projected/93940ef4-2998-4942-9c46-68e92af94436-kube-api-access-h8wc7\") pod \"service-ca-operator-777779d784-k7pr2\" (UID: \"93940ef4-2998-4942-9c46-68e92af94436\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.753371 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/be306aba-f697-4dfb-b87b-7beb769f6654-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-qn4db\" (UID: \"be306aba-f697-4dfb-b87b-7beb769f6654\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.753401 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5e602f2b-acc1-433f-852d-62aad67403f3-profile-collector-cert\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.753418 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/47911926-004f-468c-8d77-e1acefac68c3-node-bootstrap-token\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.754455 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb-cert\") pod \"ingress-canary-rtcvz\" (UID: \"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb\") " pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.756648 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhn42\" (UniqueName: \"kubernetes.io/projected/be306aba-f697-4dfb-b87b-7beb769f6654-kube-api-access-lhn42\") pod \"multus-admission-controller-857f4d67dd-qn4db\" (UID: \"be306aba-f697-4dfb-b87b-7beb769f6654\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.757092 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kssc7\" (UniqueName: \"kubernetes.io/projected/ed43fec1-dba7-4db3-bb66-f161c4c9c622-kube-api-access-kssc7\") pod \"dns-default-m42c5\" (UID: \"ed43fec1-dba7-4db3-bb66-f161c4c9c622\") " pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.759912 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.761541 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3439565e-e37c-4962-a269-ddb8a48b3f43-serving-cert\") pod \"apiserver-7bbb656c7d-gvw5z\" (UID: \"3439565e-e37c-4962-a269-ddb8a48b3f43\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.764704 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dpmn\" (UniqueName: \"kubernetes.io/projected/4777f215-78ef-47d4-be38-0e2a750168fe-kube-api-access-5dpmn\") pod \"collect-profiles-29416950-j8mc6\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.773202 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36d00e05-6326-4f98-add6-ca019b7791f4-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-lfvx7\" (UID: \"36d00e05-6326-4f98-add6-ca019b7791f4\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.776104 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-248c9\" (UniqueName: \"kubernetes.io/projected/f9063a39-025a-4fac-a196-eed330238262-kube-api-access-248c9\") pod \"package-server-manager-789f6589d5-pl9h6\" (UID: \"f9063a39-025a-4fac-a196-eed330238262\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.781086 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.799656 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.802193 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.802225 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-7v92j"] Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.814711 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qx7\" (UniqueName: \"kubernetes.io/projected/31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d-kube-api-access-l9qx7\") pod \"openshift-controller-manager-operator-756b6f6bc6-2vvw4\" (UID: \"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.819204 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.833421 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.834567 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.841649 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.842012 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.341998461 +0000 UTC m=+139.185429900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.867197 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.868652 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-m42c5" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.871185 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds77q\" (UniqueName: \"kubernetes.io/projected/460d0786-6a9e-4d3e-ae41-2cb110d42336-kube-api-access-ds77q\") pod \"openshift-apiserver-operator-796bbdcf4f-r5lkp\" (UID: \"460d0786-6a9e-4d3e-ae41-2cb110d42336\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.884984 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.925434 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.946436 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.948900 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.448875229 +0000 UTC m=+139.292306668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.949410 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/73a80455-9c8e-4061-bf7b-6afb42555cd6-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.950152 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.954160 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3d011c13-dd3f-4b33-b250-cfbc79a8e1c1-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-drjts\" (UID: \"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:53 crc kubenswrapper[4678]: E1206 10:38:53.958034 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.458015069 +0000 UTC m=+139.301446508 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.958072 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4kgq\" (UniqueName: \"kubernetes.io/projected/3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656-kube-api-access-w4kgq\") pod \"machine-config-controller-84d6567774-zpjpv\" (UID: \"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.964222 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:38:53 crc kubenswrapper[4678]: I1206 10:38:53.980194 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wng74\" (UniqueName: \"kubernetes.io/projected/72675578-a2c8-4f7b-a36b-99c5996703b5-kube-api-access-wng74\") pod \"marketplace-operator-79b997595-wbcwh\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:53.993565 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8xsx\" (UniqueName: \"kubernetes.io/projected/1262ea99-1eb6-43e0-88a2-6ecdaf79c77a-kube-api-access-m8xsx\") pod \"etcd-operator-b45778765-7b5fs\" (UID: \"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:54 crc kubenswrapper[4678]: W1206 10:38:53.998096 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5b1144c_a578_4781_b54b_9971cc41fbe6.slice/crio-717b1562bbf0cfff4de1c8c7caf65e986dad8204c5645811ce3d51a2bca02ed0 WatchSource:0}: Error finding container 717b1562bbf0cfff4de1c8c7caf65e986dad8204c5645811ce3d51a2bca02ed0: Status 404 returned error can't find the container with id 717b1562bbf0cfff4de1c8c7caf65e986dad8204c5645811ce3d51a2bca02ed0 Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:53.999790 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckthp\" (UniqueName: \"kubernetes.io/projected/5e602f2b-acc1-433f-852d-62aad67403f3-kube-api-access-ckthp\") pod \"catalog-operator-68c6474976-6ghkj\" (UID: \"5e602f2b-acc1-433f-852d-62aad67403f3\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.007937 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.024951 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.031644 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkbpf\" (UniqueName: \"kubernetes.io/projected/73a80455-9c8e-4061-bf7b-6afb42555cd6-kube-api-access-lkbpf\") pod \"ingress-operator-5b745b69d9-9zc4n\" (UID: \"73a80455-9c8e-4061-bf7b-6afb42555cd6\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.036023 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.045716 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdkpg\" (UniqueName: \"kubernetes.io/projected/1438bcb1-f3fb-4979-91ba-2bce2952edb9-kube-api-access-pdkpg\") pod \"console-operator-58897d9998-lff5h\" (UID: \"1438bcb1-f3fb-4979-91ba-2bce2952edb9\") " pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.047710 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-895zv\" (UniqueName: \"kubernetes.io/projected/6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb-kube-api-access-895zv\") pod \"ingress-canary-rtcvz\" (UID: \"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb\") " pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.051600 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.061136 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.061371 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.561335235 +0000 UTC m=+139.404766674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.062903 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmhdn\" (UniqueName: \"kubernetes.io/projected/9a6d7574-8808-4c7d-aef1-9df452c83f27-kube-api-access-xmhdn\") pod \"kube-storage-version-migrator-operator-b67b599dd-fkznw\" (UID: \"9a6d7574-8808-4c7d-aef1-9df452c83f27\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.069319 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.070120 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqgxs\" (UniqueName: \"kubernetes.io/projected/b49b0624-3162-4a68-b636-cd57545d7016-kube-api-access-cqgxs\") pod \"olm-operator-6b444d44fb-lx959\" (UID: \"b49b0624-3162-4a68-b636-cd57545d7016\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.071419 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc86x\" (UniqueName: \"kubernetes.io/projected/47911926-004f-468c-8d77-e1acefac68c3-kube-api-access-dc86x\") pod \"machine-config-server-g8tzz\" (UID: \"47911926-004f-468c-8d77-e1acefac68c3\") " pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.086615 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvhww\" (UniqueName: \"kubernetes.io/projected/f13f288c-c8dc-4df8-b89b-c325d088b30b-kube-api-access-lvhww\") pod \"console-f9d7485db-ll9jl\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.087562 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.102377 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.112289 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.148948 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-g8tzz" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.159283 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.160147 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.160757 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.660743507 +0000 UTC m=+139.504174946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.165037 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rtcvz" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.228997 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.233051 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.246926 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-4scxr"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.261553 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.261913 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.761892726 +0000 UTC m=+139.605324165 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.271543 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.291957 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.345214 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.350168 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rth98"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.352027 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.364133 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.364560 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.864544897 +0000 UTC m=+139.707976336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.367800 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.469521 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.469872 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:54.969851089 +0000 UTC m=+139.813282528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.488303 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.510890 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4scxr" event={"ID":"08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4","Type":"ContainerStarted","Data":"0f5856559842a46705e5f9e8b4c6c782bbd6dde2fe692a080c38d8abb8d6144f"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.513769 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" event={"ID":"421d8420-5eec-4a2e-9bac-f508328fdb84","Type":"ContainerStarted","Data":"1d93e180eb2a1c2b2d8076e06ab21aa3730af0029e7a68be4e944b1acc25173d"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.513855 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" event={"ID":"421d8420-5eec-4a2e-9bac-f508328fdb84","Type":"ContainerStarted","Data":"f310de66174eebd786bb5ae08962d99bc12cc0e4941aa8902109f12f915f5a5f"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.527416 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" event={"ID":"13bf0b82-57c8-46a0-99e2-cc6a975168b8","Type":"ContainerStarted","Data":"bb4b50dde98c7b386b4457168b7538828843f55ca0f43ac38018b4d5ff0952ba"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.529376 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" event={"ID":"b13d6d52-897c-4e08-ad3f-6150130d5f9f","Type":"ContainerStarted","Data":"397eccd40f364eb6b8971091d7f08d285a294b36b4048e8397d15609d85a65a0"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.530467 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-z4j5s" event={"ID":"40b9fa22-f2c6-4fb6-993a-65b61c12fed7","Type":"ContainerStarted","Data":"3c4187c7bb622ee956b0fc527d940d6924230a95a9b0d57c4bf14fa3e2689d5e"} Dec 06 10:38:54 crc kubenswrapper[4678]: W1206 10:38:54.532378 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b0941dc_500a_4284_9cc4_66b250feee93.slice/crio-6ab0f7822aaa01bcc57f3b92fc3535c387a1bf6221e799cefc9a27d0e5b155b8 WatchSource:0}: Error finding container 6ab0f7822aaa01bcc57f3b92fc3535c387a1bf6221e799cefc9a27d0e5b155b8: Status 404 returned error can't find the container with id 6ab0f7822aaa01bcc57f3b92fc3535c387a1bf6221e799cefc9a27d0e5b155b8 Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.536979 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" event={"ID":"d500b1dd-9d43-49bf-923a-c671bb6543d3","Type":"ContainerStarted","Data":"4638207a076d5a7ca01ba0efb3fc112f5064657bca9adc82f4e0ea14f56afe90"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.572807 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.576590 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.076567905 +0000 UTC m=+139.919999344 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.591215 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" event={"ID":"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732","Type":"ContainerStarted","Data":"b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.593212 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.603418 4678 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qbk7d container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.603502 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" podUID="c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.623618 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" event={"ID":"0061f615-2fb9-4e17-8ba5-62d97e8070c8","Type":"ContainerStarted","Data":"d57cde56c08b77e1cc96d4c52bfac0508f616c4f71232339cd903e385d8c6ac2"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.628148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" event={"ID":"ae8d3b71-e129-4eeb-b72c-617aa008538a","Type":"ContainerStarted","Data":"e4bd470947cbe5524642273257f173f925ee7590ecce8ebc3837a55bf7b39127"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.629990 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" event={"ID":"f5b1144c-a578-4781-b54b-9971cc41fbe6","Type":"ContainerStarted","Data":"717b1562bbf0cfff4de1c8c7caf65e986dad8204c5645811ce3d51a2bca02ed0"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.631233 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" event={"ID":"7428bf8a-2963-4011-97e0-c9f55248194b","Type":"ContainerStarted","Data":"c35757d06ae5598c93d0c89efdc2e2a7148d7a9eda58dc8bc7db8a6015a55764"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.637345 4678 generic.go:334] "Generic (PLEG): container finished" podID="e1cd25ff-72f0-414f-a4e6-51c0d5eb5662" containerID="ed70370666fc61d06a79c3f7748bc5c2175e9022b20a4ff1f2cefeeaed947bb3" exitCode=0 Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.637584 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" event={"ID":"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662","Type":"ContainerDied","Data":"ed70370666fc61d06a79c3f7748bc5c2175e9022b20a4ff1f2cefeeaed947bb3"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.637663 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" event={"ID":"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662","Type":"ContainerStarted","Data":"b631bb7306c7b8c13c47f229beec4fc0d454431962584b90dcb4246be63147bd"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.638332 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" podStartSLOduration=121.638309978 podStartE2EDuration="2m1.638309978s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:54.637461981 +0000 UTC m=+139.480893420" watchObservedRunningTime="2025-12-06 10:38:54.638309978 +0000 UTC m=+139.481741417" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.644747 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" event={"ID":"4280a0ae-d3d2-463a-9876-86967c99d560","Type":"ContainerStarted","Data":"3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8"} Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.645787 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.695093 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.695197 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.1951772 +0000 UTC m=+140.038608639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.695653 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:54 crc kubenswrapper[4678]: W1206 10:38:54.698745 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cf70a61_94e0_4b41_9823_f43bb83f2dcf.slice/crio-90eaf726e13c0b9d7b3088f8d28914a48d70f8fa30a82635a2dd6fbe98a93b67 WatchSource:0}: Error finding container 90eaf726e13c0b9d7b3088f8d28914a48d70f8fa30a82635a2dd6fbe98a93b67: Status 404 returned error can't find the container with id 90eaf726e13c0b9d7b3088f8d28914a48d70f8fa30a82635a2dd6fbe98a93b67 Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.714558 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.214520654 +0000 UTC m=+140.057952093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.781266 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-rpv8w" podStartSLOduration=121.781247125 podStartE2EDuration="2m1.781247125s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:54.723695877 +0000 UTC m=+139.567127316" watchObservedRunningTime="2025-12-06 10:38:54.781247125 +0000 UTC m=+139.624678564" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.800352 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.802461 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.302438765 +0000 UTC m=+140.145870204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.849116 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" podStartSLOduration=120.849069291 podStartE2EDuration="2m0.849069291s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:54.832129181 +0000 UTC m=+139.675560620" watchObservedRunningTime="2025-12-06 10:38:54.849069291 +0000 UTC m=+139.692500730" Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.901657 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:54 crc kubenswrapper[4678]: E1206 10:38:54.902353 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.402338389 +0000 UTC m=+140.245769828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.919630 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.926459 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-qn4db"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.934378 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6"] Dec 06 10:38:54 crc kubenswrapper[4678]: I1206 10:38:54.967509 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.003058 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.003462 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.503439245 +0000 UTC m=+140.346870684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.046050 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.095597 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.104586 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.104980 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.60496414 +0000 UTC m=+140.448395579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.215245 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.215694 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.715669957 +0000 UTC m=+140.559101396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.275467 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-rrj9g"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.303777 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.317166 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.317539 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.817526105 +0000 UTC m=+140.660957544 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.428085 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.428524 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:55.928502911 +0000 UTC m=+140.771934350 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.522515 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-m42c5"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.533922 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.534437 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.034414773 +0000 UTC m=+140.877846212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: W1206 10:38:55.590739 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47911926_004f_468c_8d77_e1acefac68c3.slice/crio-2915bfb45b49237760e3fdf84c6c2c262ea5f2b637acdf091361822c3edf1f5b WatchSource:0}: Error finding container 2915bfb45b49237760e3fdf84c6c2c262ea5f2b637acdf091361822c3edf1f5b: Status 404 returned error can't find the container with id 2915bfb45b49237760e3fdf84c6c2c262ea5f2b637acdf091361822c3edf1f5b Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.638441 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.638828 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.138804493 +0000 UTC m=+140.982235932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.708914 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.745936 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.746387 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.246370108 +0000 UTC m=+141.089801547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.846853 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.847218 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.347198505 +0000 UTC m=+141.190629944 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.888719 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2wz6s"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.937293 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4"] Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.949842 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:55 crc kubenswrapper[4678]: E1206 10:38:55.950433 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.450398163 +0000 UTC m=+141.293829672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.953113 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" event={"ID":"be306aba-f697-4dfb-b87b-7beb769f6654","Type":"ContainerStarted","Data":"70175825c3980dc5141a7b45d6d0fa1da0a6585dfaa3b25c5648b68c7ddb4625"} Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.954250 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" event={"ID":"085d0a79-150c-45f0-993a-94c8b9d1811f","Type":"ContainerStarted","Data":"7d6164c69dd2677fbf6e29871c1a505ffeb983bb234cbc7d7032a5834c590335"} Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.961889 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" event={"ID":"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9","Type":"ContainerStarted","Data":"e95921f74873e5768e4387b56ad480dce5590492c1b6edc5790420fbe55903dd"} Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.975820 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" event={"ID":"93940ef4-2998-4942-9c46-68e92af94436","Type":"ContainerStarted","Data":"4a80852f72eff53b7462c542d8527bb2a0ba44752366e841fe7e23b07c549192"} Dec 06 10:38:55 crc kubenswrapper[4678]: I1206 10:38:55.980029 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m42c5" event={"ID":"ed43fec1-dba7-4db3-bb66-f161c4c9c622","Type":"ContainerStarted","Data":"b5b41811366a4787cc61807346b4491996a0d0839178a6e55b91f02153a8cc2c"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.002635 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wbcwh"] Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.008233 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv"] Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.009994 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g8tzz" event={"ID":"47911926-004f-468c-8d77-e1acefac68c3","Type":"ContainerStarted","Data":"2915bfb45b49237760e3fdf84c6c2c262ea5f2b637acdf091361822c3edf1f5b"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.054469 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.054902 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.554881188 +0000 UTC m=+141.398312627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.070327 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" event={"ID":"2cf70a61-94e0-4b41-9823-f43bb83f2dcf","Type":"ContainerStarted","Data":"90eaf726e13c0b9d7b3088f8d28914a48d70f8fa30a82635a2dd6fbe98a93b67"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.081328 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" event={"ID":"9b0941dc-500a-4284-9cc4-66b250feee93","Type":"ContainerStarted","Data":"6ab0f7822aaa01bcc57f3b92fc3535c387a1bf6221e799cefc9a27d0e5b155b8"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.090804 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z"] Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.103429 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" event={"ID":"4777f215-78ef-47d4-be38-0e2a750168fe","Type":"ContainerStarted","Data":"9ce49b55526ec52286791f13b14dadc3c172a85563dfa4471ee094a9072b7ef2"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.151321 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" event={"ID":"53361e7d-fe05-4f40-a442-307cb30ecd1c","Type":"ContainerStarted","Data":"13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.151379 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.157632 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.158077 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.658063586 +0000 UTC m=+141.501495025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.222390 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-z4j5s" event={"ID":"40b9fa22-f2c6-4fb6-993a-65b61c12fed7","Type":"ContainerStarted","Data":"e824fc5275fdec81a252277181dd9416799ae3498f86334b37242ae94f2520b0"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.258706 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.261961 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.761931299 +0000 UTC m=+141.605362728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.267299 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7"] Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.267347 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" event={"ID":"f9063a39-025a-4fac-a196-eed330238262","Type":"ContainerStarted","Data":"af7e2caf18f86def4e8e1705181207e05eb863c81adfdf0bc5c458d1e9e71aef"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.323383 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-z4j5s" podStartSLOduration=122.323359741 podStartE2EDuration="2m2.323359741s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:56.282311793 +0000 UTC m=+141.125743232" watchObservedRunningTime="2025-12-06 10:38:56.323359741 +0000 UTC m=+141.166791180" Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.332363 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" podStartSLOduration=123.33233109 podStartE2EDuration="2m3.33233109s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:56.308508053 +0000 UTC m=+141.151939492" watchObservedRunningTime="2025-12-06 10:38:56.33233109 +0000 UTC m=+141.175762529" Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.367269 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.369153 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.86913613 +0000 UTC m=+141.712567569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.403276 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.413120 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:38:56 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:38:56 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:38:56 crc kubenswrapper[4678]: healthz check failed Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.413187 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.436996 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" event={"ID":"0bbdcc6e-f0b1-4541-877c-9bec206a2897","Type":"ContainerStarted","Data":"45fc7a479b8254941796fbab5d123933d26047a81b7c3512bc99be6263821b27"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.450146 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" event={"ID":"460d0786-6a9e-4d3e-ae41-2cb110d42336","Type":"ContainerStarted","Data":"fab55c53c4390f016ca8433a0bb79d2f11085c1fd58a37610c46da596f036cbe"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.470562 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.471178 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:56.971153468 +0000 UTC m=+141.814584907 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.573347 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" event={"ID":"d500b1dd-9d43-49bf-923a-c671bb6543d3","Type":"ContainerStarted","Data":"19c38cef3ddeeea66b78c9ba9e0ef3bdd15bd32786dcc32e6ab20bc48ea9dfa3"} Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.573864 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.574282 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.07426482 +0000 UTC m=+141.917696259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.587285 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.637400 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-h8wxt" podStartSLOduration=123.637376426 podStartE2EDuration="2m3.637376426s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:56.624941642 +0000 UTC m=+141.468373071" watchObservedRunningTime="2025-12-06 10:38:56.637376426 +0000 UTC m=+141.480807865" Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.698328 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.699846 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.199825087 +0000 UTC m=+142.043256526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.782478 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj"] Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.802512 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.802906 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.302892596 +0000 UTC m=+142.146324035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.881390 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rtcvz"] Dec 06 10:38:56 crc kubenswrapper[4678]: I1206 10:38:56.904777 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:56 crc kubenswrapper[4678]: E1206 10:38:56.905310 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.40528575 +0000 UTC m=+142.248717189 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.008143 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.008783 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.508761787 +0000 UTC m=+142.352193226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.024783 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-lff5h"] Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.111754 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.112192 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.61217499 +0000 UTC m=+142.455606429 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.140862 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7b5fs"] Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.151357 4678 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5vk2z container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.9:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.151462 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" podUID="53361e7d-fe05-4f40-a442-307cb30ecd1c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.9:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.222537 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.223458 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.723443986 +0000 UTC m=+142.566875415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.246797 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959"] Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.249562 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw"] Dec 06 10:38:57 crc kubenswrapper[4678]: W1206 10:38:57.280446 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1438bcb1_f3fb_4979_91ba_2bce2952edb9.slice/crio-8107d8dcaa8c767c2135e05bb3174d1f971312e260c86eec1ec0dbeaebcfd6f8 WatchSource:0}: Error finding container 8107d8dcaa8c767c2135e05bb3174d1f971312e260c86eec1ec0dbeaebcfd6f8: Status 404 returned error can't find the container with id 8107d8dcaa8c767c2135e05bb3174d1f971312e260c86eec1ec0dbeaebcfd6f8 Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.334289 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.334997 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.83497182 +0000 UTC m=+142.678403259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.364894 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n"] Dec 06 10:38:57 crc kubenswrapper[4678]: W1206 10:38:57.402067 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a6d7574_8808_4c7d_aef1_9df452c83f27.slice/crio-749f0cef69136db337955fc8b97da37d0c522e8fd7a375f6a4157a1234ad9642 WatchSource:0}: Error finding container 749f0cef69136db337955fc8b97da37d0c522e8fd7a375f6a4157a1234ad9642: Status 404 returned error can't find the container with id 749f0cef69136db337955fc8b97da37d0c522e8fd7a375f6a4157a1234ad9642 Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.421040 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:38:57 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:38:57 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:38:57 crc kubenswrapper[4678]: healthz check failed Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.421116 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.437338 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.437726 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:57.937711067 +0000 UTC m=+142.781142506 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.545203 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ll9jl"] Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.548141 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.548624 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.048602848 +0000 UTC m=+142.892034287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.620835 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lff5h" event={"ID":"1438bcb1-f3fb-4979-91ba-2bce2952edb9","Type":"ContainerStarted","Data":"8107d8dcaa8c767c2135e05bb3174d1f971312e260c86eec1ec0dbeaebcfd6f8"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.651014 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.651395 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.151379327 +0000 UTC m=+142.994810766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.655714 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" event={"ID":"421d8420-5eec-4a2e-9bac-f508328fdb84","Type":"ContainerStarted","Data":"6ed97ed668d292030fe6fee7b28205ef6fe73029f26a81d761c02daf96b8a643"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.690847 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" event={"ID":"9a6d7574-8808-4c7d-aef1-9df452c83f27","Type":"ContainerStarted","Data":"749f0cef69136db337955fc8b97da37d0c522e8fd7a375f6a4157a1234ad9642"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.710103 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r85v9" podStartSLOduration=124.710078454 podStartE2EDuration="2m4.710078454s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:57.707425024 +0000 UTC m=+142.550856463" watchObservedRunningTime="2025-12-06 10:38:57.710078454 +0000 UTC m=+142.553509883" Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.711912 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" event={"ID":"72675578-a2c8-4f7b-a36b-99c5996703b5","Type":"ContainerStarted","Data":"e7d2013468570424d777b54785cde7d1a85eb5578d2e3dfdfd672e2c7f3ca7bf"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.753157 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.754636 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" event={"ID":"36d00e05-6326-4f98-add6-ca019b7791f4","Type":"ContainerStarted","Data":"bb95e2a9b7b48efd153202477384f16100ea056b2966817895844386e065ae77"} Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.754750 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.254730396 +0000 UTC m=+143.098161835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.835897 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" event={"ID":"2cf70a61-94e0-4b41-9823-f43bb83f2dcf","Type":"ContainerStarted","Data":"d2f1bedec9dd17f66c55cfc5fc85db04c1020f43e7145d3c900eb0c147d0282f"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.876425 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.877108 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.377078514 +0000 UTC m=+143.220509953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.877408 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" event={"ID":"7428bf8a-2963-4011-97e0-c9f55248194b","Type":"ContainerStarted","Data":"59e9e651dc6179cb6d62744c1455b9ed49a89193ce6752a13e8905235af0d2b8"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.920870 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" event={"ID":"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1","Type":"ContainerStarted","Data":"1907e1e810c16507ba9d26e5db05f8bfd00fd8c7e79e2208acaeb9bc501a6fa4"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.947822 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-f6m8w" podStartSLOduration=123.947799767 podStartE2EDuration="2m3.947799767s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:57.946559804 +0000 UTC m=+142.789991243" watchObservedRunningTime="2025-12-06 10:38:57.947799767 +0000 UTC m=+142.791231226" Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.969871 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" event={"ID":"0061f615-2fb9-4e17-8ba5-62d97e8070c8","Type":"ContainerStarted","Data":"da3ebe483aec2156736ed82eabe595dab5527bf4fd43bb15736bd44b4c71a0b1"} Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.981741 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.981806 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.481787216 +0000 UTC m=+143.325218655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:57 crc kubenswrapper[4678]: I1206 10:38:57.989970 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:57 crc kubenswrapper[4678]: E1206 10:38:57.991384 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.491369627 +0000 UTC m=+143.334801066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.017085 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" event={"ID":"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656","Type":"ContainerStarted","Data":"d8d8f9f90f8a80470ac0e70f669e6e26829172d300ef247a0aee57d34dcfd226"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.018508 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-r5rng" podStartSLOduration=124.018443295 podStartE2EDuration="2m4.018443295s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:58.011208944 +0000 UTC m=+142.854640393" watchObservedRunningTime="2025-12-06 10:38:58.018443295 +0000 UTC m=+142.861874734" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.041530 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" event={"ID":"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d","Type":"ContainerStarted","Data":"0a6a6a322a5bcbf68dc81111ed5ef3cebdb9b1e2a8f349458783d796fce13c95"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.085350 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" event={"ID":"9b0941dc-500a-4284-9cc4-66b250feee93","Type":"ContainerStarted","Data":"41f03a88626513eba5dfb4bafa02b4220708b693a33312ed5cc1cbe976a59eea"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.100843 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.102263 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.602241376 +0000 UTC m=+143.445672815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.130969 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" event={"ID":"5e602f2b-acc1-433f-852d-62aad67403f3","Type":"ContainerStarted","Data":"e4b4f055e44375456d254651036140ca8354616e0eb0d9e3b7b9e8c2bcbc9448"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.136237 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" event={"ID":"f5b1144c-a578-4781-b54b-9971cc41fbe6","Type":"ContainerStarted","Data":"c1762747279b229f776877e2e379a6b1dd466dbe8b4936971702de794de30524"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.157606 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" event={"ID":"b13d6d52-897c-4e08-ad3f-6150130d5f9f","Type":"ContainerStarted","Data":"206a63536c51aac41bbdc615d32a4e8fa28e31b40eccd830c55c5e90d5424c53"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.158934 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" event={"ID":"f9063a39-025a-4fac-a196-eed330238262","Type":"ContainerStarted","Data":"4a2b648342c059d9ec1540e1103f98befcd2fe3372933584cec44920d8550cb0"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.159691 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" event={"ID":"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a","Type":"ContainerStarted","Data":"32ef60aa8993fef37afe409025bea49f1eb4a2feb31fc9ab86003d8ef4e51444"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.168471 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" event={"ID":"b49b0624-3162-4a68-b636-cd57545d7016","Type":"ContainerStarted","Data":"26fd9def01ae609adf40241106425b44ac1c4b9e90ce311ffb9e6a1b3a957dee"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.171823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" event={"ID":"460d0786-6a9e-4d3e-ae41-2cb110d42336","Type":"ContainerStarted","Data":"ed16f2ebdd859425c8207c6c68cdba0926982e901f456f6b9272040f7e097d7c"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.176256 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" event={"ID":"73a80455-9c8e-4061-bf7b-6afb42555cd6","Type":"ContainerStarted","Data":"ace403b4fab562ea293bbe693e465a6539ab40d2fb53c5c1b4306419f5f2f697"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.205320 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.206331 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.706296752 +0000 UTC m=+143.549728191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.213940 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-r5lkp" podStartSLOduration=125.213914479 podStartE2EDuration="2m5.213914479s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:58.213800522 +0000 UTC m=+143.057231961" watchObservedRunningTime="2025-12-06 10:38:58.213914479 +0000 UTC m=+143.057345918" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.215420 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dvssp" podStartSLOduration=124.215413542 podStartE2EDuration="2m4.215413542s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:58.125018733 +0000 UTC m=+142.968450172" watchObservedRunningTime="2025-12-06 10:38:58.215413542 +0000 UTC m=+143.058844981" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.273049 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-4scxr" event={"ID":"08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4","Type":"ContainerStarted","Data":"9012b3c2779af7701cf8ea87c1db9034fce013c9cbcd379a18db9b8c86389a3b"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.274439 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-4scxr" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.295750 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-4scxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.295835 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4scxr" podUID="08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.306969 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.308146 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.808120178 +0000 UTC m=+143.651551617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.321885 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" event={"ID":"ae8d3b71-e129-4eeb-b72c-617aa008538a","Type":"ContainerStarted","Data":"f7b37a0404c91743247506e080291a7ad0068e84baa823494aa55b0ca50f120e"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.343059 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-4scxr" podStartSLOduration=125.343034738 podStartE2EDuration="2m5.343034738s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:58.342421656 +0000 UTC m=+143.185853095" watchObservedRunningTime="2025-12-06 10:38:58.343034738 +0000 UTC m=+143.186466177" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.359007 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" event={"ID":"93940ef4-2998-4942-9c46-68e92af94436","Type":"ContainerStarted","Data":"8f7b47b342d5a1eba0488dc3615e79d3cebfb7d55e3422a4539c7ea36fdee0b2"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.372263 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" event={"ID":"3439565e-e37c-4962-a269-ddb8a48b3f43","Type":"ContainerStarted","Data":"beaec55e6f53b0d7bc469b8cf4207931827210f520123ee17f4eb97c0feea3b6"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.407285 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:38:58 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:38:58 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:38:58 crc kubenswrapper[4678]: healthz check failed Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.407774 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.416996 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.419164 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:58.919140477 +0000 UTC m=+143.762571916 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.427669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" event={"ID":"a0c46190-8e1a-4068-ba9f-9d3751787308","Type":"ContainerStarted","Data":"526c91ee3c55a376d76808cfb17ddae4f706962c5afcbdfdd901ccd869c897ca"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.457465 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-lfrsj" podStartSLOduration=125.457441367 podStartE2EDuration="2m5.457441367s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:58.390138597 +0000 UTC m=+143.233570036" watchObservedRunningTime="2025-12-06 10:38:58.457441367 +0000 UTC m=+143.300872806" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.458329 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-k7pr2" podStartSLOduration=124.458323128 podStartE2EDuration="2m4.458323128s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:58.457705105 +0000 UTC m=+143.301136544" watchObservedRunningTime="2025-12-06 10:38:58.458323128 +0000 UTC m=+143.301754567" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.464719 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" event={"ID":"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9","Type":"ContainerStarted","Data":"2154e052f6135249bea8ebd6700c6530e801f7067472d5159b0cdf5dab9ed92d"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.501619 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rtcvz" event={"ID":"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb","Type":"ContainerStarted","Data":"6770fa45df06cd4129f234114434bbcdd633a4d40da5e03657260083eb41c92b"} Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.509898 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.517918 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.519297 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.019277727 +0000 UTC m=+143.862709166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.628913 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.635538 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.135519731 +0000 UTC m=+143.978951170 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.732249 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.732684 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.232664038 +0000 UTC m=+144.076095477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.846735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.847983 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.347964929 +0000 UTC m=+144.191396368 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:58 crc kubenswrapper[4678]: I1206 10:38:58.953319 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:58 crc kubenswrapper[4678]: E1206 10:38:58.953854 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.453832688 +0000 UTC m=+144.297264127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.057910 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.058352 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.558334865 +0000 UTC m=+144.401766304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.159233 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.160191 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.660152869 +0000 UTC m=+144.503584318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.262342 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.262965 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.762945749 +0000 UTC m=+144.606377188 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.363388 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.363975 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.863941278 +0000 UTC m=+144.707372717 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.411352 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:38:59 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:38:59 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:38:59 crc kubenswrapper[4678]: healthz check failed Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.411417 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.466476 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.466936 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:38:59.966915072 +0000 UTC m=+144.810346501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.506139 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.506203 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.568078 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.568540 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.068517301 +0000 UTC m=+144.911948740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.595183 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" event={"ID":"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656","Type":"ContainerStarted","Data":"bcd7b33ab93cf97b06843f683e5332b673e51301b35b0451c0ea4d95bb9d4975"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.595239 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" event={"ID":"3dedcfb5-5efb-4d5e-9f28-d9f3ea84e656","Type":"ContainerStarted","Data":"43c20334b9c53d0112650ab574cf466bf9dab4d1ddabb0b557e9f5c4b93c5abc"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.613385 4678 generic.go:334] "Generic (PLEG): container finished" podID="6508c5a9-1514-4fe7-88cf-9d5f03af9ff9" containerID="2154e052f6135249bea8ebd6700c6530e801f7067472d5159b0cdf5dab9ed92d" exitCode=0 Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.613506 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" event={"ID":"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9","Type":"ContainerDied","Data":"2154e052f6135249bea8ebd6700c6530e801f7067472d5159b0cdf5dab9ed92d"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.617608 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" event={"ID":"b13d6d52-897c-4e08-ad3f-6150130d5f9f","Type":"ContainerStarted","Data":"61abf5bf333aa87585e1433fb8cafa8da5f71be42926301ff50435779624966e"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.631318 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" event={"ID":"5e602f2b-acc1-433f-852d-62aad67403f3","Type":"ContainerStarted","Data":"767d046fc8261a61240484c10ebe4b369603475eb577c700fbe0869dfe2c2745"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.632647 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.655140 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-g8tzz" event={"ID":"47911926-004f-468c-8d77-e1acefac68c3","Type":"ContainerStarted","Data":"af36b6256ac88419630505af352cf9fa25757a6a8a9b27024dd8021467b07634"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.669917 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.670439 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.170422352 +0000 UTC m=+145.013853791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.670708 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" event={"ID":"0bbdcc6e-f0b1-4541-877c-9bec206a2897","Type":"ContainerStarted","Data":"08b3c0e24a090f69a510008a8a8243fe1634e6fe63bd0c40f765f06bcd7e1adc"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.670880 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.683111 4678 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-ksqhs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.683240 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" podUID="0bbdcc6e-f0b1-4541-877c-9bec206a2897" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.701641 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-p79pr" podStartSLOduration=125.70161119 podStartE2EDuration="2m5.70161119s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:59.668603089 +0000 UTC m=+144.512034548" watchObservedRunningTime="2025-12-06 10:38:59.70161119 +0000 UTC m=+144.545042629" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.702153 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-g8tzz" podStartSLOduration=9.702148267 podStartE2EDuration="9.702148267s" podCreationTimestamp="2025-12-06 10:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:59.700477923 +0000 UTC m=+144.543909362" watchObservedRunningTime="2025-12-06 10:38:59.702148267 +0000 UTC m=+144.545579706" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.719255 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" event={"ID":"085d0a79-150c-45f0-993a-94c8b9d1811f","Type":"ContainerStarted","Data":"b940a1061da1b1e7ad78a2361a1b60ef1199daaeabf54018bd845eac1cbb29ca"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.752213 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" event={"ID":"9a6d7574-8808-4c7d-aef1-9df452c83f27","Type":"ContainerStarted","Data":"038877c78c4c5b9b90016d30fe0d58f0932cd48c8ee614dac73473e526fb55f6"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.752725 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" podStartSLOduration=125.75269792 podStartE2EDuration="2m5.75269792s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:59.752476084 +0000 UTC m=+144.595907523" watchObservedRunningTime="2025-12-06 10:38:59.75269792 +0000 UTC m=+144.596129359" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.773117 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.774828 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.27479067 +0000 UTC m=+145.118222109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.791854 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" event={"ID":"72675578-a2c8-4f7b-a36b-99c5996703b5","Type":"ContainerStarted","Data":"7e86c9e1dbf04cb5622d86c7e5f2fc46f0253c4be77b8f0ad683cf5210a7dd58"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.792776 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.798606 4678 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wbcwh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.798645 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.832408 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" podStartSLOduration=125.8323802 podStartE2EDuration="2m5.8323802s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:59.820433169 +0000 UTC m=+144.663864608" watchObservedRunningTime="2025-12-06 10:38:59.8323802 +0000 UTC m=+144.675811639" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.835235 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" event={"ID":"36d00e05-6326-4f98-add6-ca019b7791f4","Type":"ContainerStarted","Data":"02acb01e84b0ff19cd4072f2151d70a4afef6580de22a7638c38c8e5cd7edc1d"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.845802 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" event={"ID":"f9063a39-025a-4fac-a196-eed330238262","Type":"ContainerStarted","Data":"35f51e7b0eb489029917bf3efee25cca522f7692ccc3a628fb750fd70d0d708e"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.846366 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.869997 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" event={"ID":"2cf70a61-94e0-4b41-9823-f43bb83f2dcf","Type":"ContainerStarted","Data":"ccb0857d06b35f62fd95114d30813988934494f3233fbdbf339607c35197a275"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.873893 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6ghkj" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.876453 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.878545 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.378527135 +0000 UTC m=+145.221958574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.879407 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-rrj9g" podStartSLOduration=125.879388283 podStartE2EDuration="2m5.879388283s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:59.879377352 +0000 UTC m=+144.722808781" watchObservedRunningTime="2025-12-06 10:38:59.879388283 +0000 UTC m=+144.722819722" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.881010 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" event={"ID":"3d011c13-dd3f-4b33-b250-cfbc79a8e1c1","Type":"ContainerStarted","Data":"46a5be429429c411d62d056cd3cbeac11fbbb420cf9847990fc61af8a27865e0"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.883095 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m42c5" event={"ID":"ed43fec1-dba7-4db3-bb66-f161c4c9c622","Type":"ContainerStarted","Data":"a922de663f5198979980409de056591fc130e37a6fcb6e150d1954c219f2692d"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.904860 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-fkznw" podStartSLOduration=125.904835601 podStartE2EDuration="2m5.904835601s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:59.903699144 +0000 UTC m=+144.747130593" watchObservedRunningTime="2025-12-06 10:38:59.904835601 +0000 UTC m=+144.748267040" Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.910868 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" event={"ID":"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662","Type":"ContainerStarted","Data":"e62f959de2bf3cfd620c962d2323e6156b242f3415a26d5adbe92445db767b78"} Dec 06 10:38:59 crc kubenswrapper[4678]: I1206 10:38:59.977671 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:38:59 crc kubenswrapper[4678]: E1206 10:38:59.982833 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.482805976 +0000 UTC m=+145.326237415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:38:59.999713 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll9jl" event={"ID":"f13f288c-c8dc-4df8-b89b-c325d088b30b","Type":"ContainerStarted","Data":"532bb9d4a6fd1b7984067a6d937ac24c9b025e891f50325fbfb3997a8c59ee20"} Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.003326 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-drjts" podStartSLOduration=126.003311478 podStartE2EDuration="2m6.003311478s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:38:59.975871055 +0000 UTC m=+144.819302494" watchObservedRunningTime="2025-12-06 10:39:00.003311478 +0000 UTC m=+144.846742917" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.024669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" event={"ID":"4777f215-78ef-47d4-be38-0e2a750168fe","Type":"ContainerStarted","Data":"1ad63fcf3ea70355e95429e928b0d8ad6bb8651a83bdf2cf72bf9cc82a179b61"} Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.038481 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-lfvx7" podStartSLOduration=126.038464525 podStartE2EDuration="2m6.038464525s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.004738996 +0000 UTC m=+144.848170435" watchObservedRunningTime="2025-12-06 10:39:00.038464525 +0000 UTC m=+144.881895964" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.047913 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" event={"ID":"f5b1144c-a578-4781-b54b-9971cc41fbe6","Type":"ContainerStarted","Data":"8c6adeebce2c9df88a3175d2bad225cbda5eb9d9489aa831bd15d2a9d5710cf5"} Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.052400 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" event={"ID":"b49b0624-3162-4a68-b636-cd57545d7016","Type":"ContainerStarted","Data":"5b06eb94042666c4faa8921bd17432e74cb7a11f40ca90f6eb1b64c60e32da9d"} Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.053120 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.058581 4678 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-lx959 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.058619 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" podUID="b49b0624-3162-4a68-b636-cd57545d7016" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.073703 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-hfsx2" podStartSLOduration=126.073684278 podStartE2EDuration="2m6.073684278s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.072954428 +0000 UTC m=+144.916385887" watchObservedRunningTime="2025-12-06 10:39:00.073684278 +0000 UTC m=+144.917115717" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.087295 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.087723 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.58770727 +0000 UTC m=+145.431138709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.111646 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" event={"ID":"be306aba-f697-4dfb-b87b-7beb769f6654","Type":"ContainerStarted","Data":"92e69f08263e2d8546b50dd235384b8791487a9c93ff16c193d59e225fa5a9ab"} Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.138667 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" podStartSLOduration=126.138645379 podStartE2EDuration="2m6.138645379s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.13689292 +0000 UTC m=+144.980324349" watchObservedRunningTime="2025-12-06 10:39:00.138645379 +0000 UTC m=+144.982076818" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.167648 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" event={"ID":"31d3e08a-c47f-475c-a9ce-f7e83e9ccc5d","Type":"ContainerStarted","Data":"253be650060cef5c7301e53f8b1070407ad7e5525a5b9703cb64e92a2df59beb"} Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.169640 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-4scxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.169698 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4scxr" podUID="08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.190129 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.191214 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.691172836 +0000 UTC m=+145.534604275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.198252 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" podStartSLOduration=126.198233585 podStartE2EDuration="2m6.198233585s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.197216157 +0000 UTC m=+145.040647596" watchObservedRunningTime="2025-12-06 10:39:00.198233585 +0000 UTC m=+145.041665014" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.237602 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-7v92j" podStartSLOduration=127.237586348 podStartE2EDuration="2m7.237586348s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.234652339 +0000 UTC m=+145.078083778" watchObservedRunningTime="2025-12-06 10:39:00.237586348 +0000 UTC m=+145.081017777" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.291594 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-2vvw4" podStartSLOduration=127.291575405 podStartE2EDuration="2m7.291575405s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.285654552 +0000 UTC m=+145.129085991" watchObservedRunningTime="2025-12-06 10:39:00.291575405 +0000 UTC m=+145.135006844" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.297716 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.305046 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.805025638 +0000 UTC m=+145.648457077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.322056 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" podStartSLOduration=127.322036243 podStartE2EDuration="2m7.322036243s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.31639734 +0000 UTC m=+145.159828799" watchObservedRunningTime="2025-12-06 10:39:00.322036243 +0000 UTC m=+145.165467672" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.399054 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.399401 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:00.899380616 +0000 UTC m=+145.742812055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.406635 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:00 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:00 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:00 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.406704 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.500308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.500837 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.000814114 +0000 UTC m=+145.844245553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.601228 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.601674 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.101645091 +0000 UTC m=+145.945076530 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.703093 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.703524 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.203507659 +0000 UTC m=+146.046939098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.803958 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.804964 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.304942398 +0000 UTC m=+146.148373837 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:00 crc kubenswrapper[4678]: I1206 10:39:00.906562 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:00 crc kubenswrapper[4678]: E1206 10:39:00.906948 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.406933314 +0000 UTC m=+146.250364753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.007661 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.008021 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.508001007 +0000 UTC m=+146.351432436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.109243 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.109693 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.609670772 +0000 UTC m=+146.453102211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.169681 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" event={"ID":"1262ea99-1eb6-43e0-88a2-6ecdaf79c77a","Type":"ContainerStarted","Data":"13b6f31dc0ccf5c31c23b44ca67eab2764f70f859ef2a2b8985ed821630a837a"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.171999 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-m42c5" event={"ID":"ed43fec1-dba7-4db3-bb66-f161c4c9c622","Type":"ContainerStarted","Data":"8fae840a2169cf1868752449bc881545be54e28f9b58306b11dd868138a95f5e"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.172681 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-m42c5" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.174935 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" event={"ID":"e1cd25ff-72f0-414f-a4e6-51c0d5eb5662","Type":"ContainerStarted","Data":"a7e5b1f96380d5dd148779f8382b255fb230ea65cdde29cd01e366c2aa8c9358"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.177954 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" event={"ID":"6508c5a9-1514-4fe7-88cf-9d5f03af9ff9","Type":"ContainerStarted","Data":"9e75946633fabf1b235605037c3bd325a25fc38b20e275f2a353b5c9bc5a08db"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.178362 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.179826 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll9jl" event={"ID":"f13f288c-c8dc-4df8-b89b-c325d088b30b","Type":"ContainerStarted","Data":"fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.182262 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" event={"ID":"be306aba-f697-4dfb-b87b-7beb769f6654","Type":"ContainerStarted","Data":"5cb24a59cb571955ca2ebe96baafa3f459f3700966e361504d8efe7c522c2ce1"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.184030 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-lff5h" event={"ID":"1438bcb1-f3fb-4979-91ba-2bce2952edb9","Type":"ContainerStarted","Data":"4e9c92625e9dbc4888b9f82274c6f70c35ba900091dec81ecd934a56b3a7708b"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.184932 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.186650 4678 patch_prober.go:28] interesting pod/console-operator-58897d9998-lff5h container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.186704 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-lff5h" podUID="1438bcb1-f3fb-4979-91ba-2bce2952edb9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.186761 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" event={"ID":"a0c46190-8e1a-4068-ba9f-9d3751787308","Type":"ContainerStarted","Data":"cb55fdffa9f8f6552f4da06b7dc3aeb5ad173470025aead99e1bc40a22ce1694"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.189128 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" event={"ID":"73a80455-9c8e-4061-bf7b-6afb42555cd6","Type":"ContainerStarted","Data":"888fe76cbaa2b4c44aa1a0d808d296cbddf9615e86aa765401db501d7efa92fb"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.189160 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" event={"ID":"73a80455-9c8e-4061-bf7b-6afb42555cd6","Type":"ContainerStarted","Data":"0a9c30736bfd1b45518b0690b414336b614ea6daef34d986066352d5299dbaea"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.191086 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rtcvz" event={"ID":"6836a2f0-dc57-4c4e-9dc5-d333ab8f25bb","Type":"ContainerStarted","Data":"7821f35226e290739dba2f98a6f2160540a4fc16d6f7a172a34600780f4d4a3c"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.193348 4678 generic.go:334] "Generic (PLEG): container finished" podID="3439565e-e37c-4962-a269-ddb8a48b3f43" containerID="e67541ebe5aa76074c71270eeb29debb8c71c1c017d8b318fffda0f3d3796795" exitCode=0 Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.193434 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" event={"ID":"3439565e-e37c-4962-a269-ddb8a48b3f43","Type":"ContainerDied","Data":"e67541ebe5aa76074c71270eeb29debb8c71c1c017d8b318fffda0f3d3796795"} Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.194451 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-4scxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.194500 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4scxr" podUID="08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.194540 4678 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wbcwh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.194609 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.210418 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.210610 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.710576475 +0000 UTC m=+146.554007914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.210739 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.211096 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.711081769 +0000 UTC m=+146.554513208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.223335 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" podStartSLOduration=127.223313279 podStartE2EDuration="2m7.223313279s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:00.34099088 +0000 UTC m=+145.184422319" watchObservedRunningTime="2025-12-06 10:39:01.223313279 +0000 UTC m=+146.066744718" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.223871 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7b5fs" podStartSLOduration=128.223867957 podStartE2EDuration="2m8.223867957s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.221370248 +0000 UTC m=+146.064801687" watchObservedRunningTime="2025-12-06 10:39:01.223867957 +0000 UTC m=+146.067299396" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.257759 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ll9jl" podStartSLOduration=128.257735317 podStartE2EDuration="2m8.257735317s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.252526464 +0000 UTC m=+146.095957903" watchObservedRunningTime="2025-12-06 10:39:01.257735317 +0000 UTC m=+146.101166756" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.287644 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-ksqhs" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.311793 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.313329 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.81329965 +0000 UTC m=+146.656731089 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.334939 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-lx959" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.349127 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9zc4n" podStartSLOduration=127.349107322 podStartE2EDuration="2m7.349107322s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.348451128 +0000 UTC m=+146.191882567" watchObservedRunningTime="2025-12-06 10:39:01.349107322 +0000 UTC m=+146.192538761" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.350677 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-qn4db" podStartSLOduration=127.350672229 podStartE2EDuration="2m7.350672229s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.296821611 +0000 UTC m=+146.140253050" watchObservedRunningTime="2025-12-06 10:39:01.350672229 +0000 UTC m=+146.194103668" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.389668 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zpjpv" podStartSLOduration=127.389644505 podStartE2EDuration="2m7.389644505s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.387894956 +0000 UTC m=+146.231326405" watchObservedRunningTime="2025-12-06 10:39:01.389644505 +0000 UTC m=+146.233075954" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.405854 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:01 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:01 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:01 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.405929 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.415053 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.415546 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:01.915529203 +0000 UTC m=+146.758960642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.445850 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" podStartSLOduration=128.445827271 podStartE2EDuration="2m8.445827271s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.445238841 +0000 UTC m=+146.288670280" watchObservedRunningTime="2025-12-06 10:39:01.445827271 +0000 UTC m=+146.289258710" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.515980 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.516579 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.016553744 +0000 UTC m=+146.859985183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.518346 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" podStartSLOduration=128.518331144 podStartE2EDuration="2m8.518331144s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.499429211 +0000 UTC m=+146.342860680" watchObservedRunningTime="2025-12-06 10:39:01.518331144 +0000 UTC m=+146.361762583" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.533970 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rtcvz" podStartSLOduration=11.533951865 podStartE2EDuration="11.533951865s" podCreationTimestamp="2025-12-06 10:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.533051554 +0000 UTC m=+146.376482993" watchObservedRunningTime="2025-12-06 10:39:01.533951865 +0000 UTC m=+146.377383294" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.625208 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.625546 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.125533425 +0000 UTC m=+146.968964854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.629538 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-m42c5" podStartSLOduration=11.629516445 podStartE2EDuration="11.629516445s" podCreationTimestamp="2025-12-06 10:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.623988919 +0000 UTC m=+146.467420358" watchObservedRunningTime="2025-12-06 10:39:01.629516445 +0000 UTC m=+146.472947884" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.681959 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s8p98"] Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.686606 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.699965 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.725996 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.730716 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.230681415 +0000 UTC m=+147.074112854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.732791 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.733454 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.233437022 +0000 UTC m=+147.076868461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.770384 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s8p98"] Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.841653 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vkwz2"] Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.845001 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.845224 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-utilities\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.845272 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-catalog-content\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.845354 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r66k5\" (UniqueName: \"kubernetes.io/projected/29fca54e-1ec1-484b-b3c4-431ecaedd87b-kube-api-access-r66k5\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.845524 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.345472021 +0000 UTC m=+147.188903460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.856458 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.856668 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.856680 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.859805 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.903850 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-lff5h" podStartSLOduration=128.903829373 podStartE2EDuration="2m8.903829373s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:01.893211923 +0000 UTC m=+146.736643362" watchObservedRunningTime="2025-12-06 10:39:01.903829373 +0000 UTC m=+146.747260812" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.920466 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkwz2"] Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.952059 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zcdm\" (UniqueName: \"kubernetes.io/projected/512340e2-4c40-48fd-ac35-8f84f1de5c6a-kube-api-access-5zcdm\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.952414 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-utilities\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.957621 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-utilities\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.957758 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.957839 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-catalog-content\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.958015 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-catalog-content\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.958117 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r66k5\" (UniqueName: \"kubernetes.io/projected/29fca54e-1ec1-484b-b3c4-431ecaedd87b-kube-api-access-r66k5\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.954045 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-utilities\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:01 crc kubenswrapper[4678]: E1206 10:39:01.961875 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.461858914 +0000 UTC m=+147.305290353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:01 crc kubenswrapper[4678]: I1206 10:39:01.962268 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-catalog-content\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.000712 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r66k5\" (UniqueName: \"kubernetes.io/projected/29fca54e-1ec1-484b-b3c4-431ecaedd87b-kube-api-access-r66k5\") pod \"community-operators-s8p98\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.009277 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.037163 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5gkj5"] Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.038410 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.065087 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.065452 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.565416177 +0000 UTC m=+147.408847616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.065920 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zcdm\" (UniqueName: \"kubernetes.io/projected/512340e2-4c40-48fd-ac35-8f84f1de5c6a-kube-api-access-5zcdm\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.066138 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-utilities\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.066278 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.066443 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-catalog-content\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.066635 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-catalog-content\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.066744 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-utilities\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.066908 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8hdg\" (UniqueName: \"kubernetes.io/projected/582c80c6-20e2-4a0e-9871-f1f4db232ca2-kube-api-access-l8hdg\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.068040 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-utilities\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.068426 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.56841028 +0000 UTC m=+147.411841719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.072988 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-catalog-content\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.168940 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.169222 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-catalog-content\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.169261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-utilities\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.169282 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8hdg\" (UniqueName: \"kubernetes.io/projected/582c80c6-20e2-4a0e-9871-f1f4db232ca2-kube-api-access-l8hdg\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.169712 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.669695108 +0000 UTC m=+147.513126547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.189569 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5gkj5"] Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.242555 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" event={"ID":"a0c46190-8e1a-4068-ba9f-9d3751787308","Type":"ContainerStarted","Data":"81ff186ea4b12748b62b063bb38d1815da8a58ffd9b4aa889df46e2b833d75cb"} Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.243355 4678 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-wbcwh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" start-of-body= Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.243423 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.39:8080/healthz\": dial tcp 10.217.0.39:8080: connect: connection refused" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.252570 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zcdm\" (UniqueName: \"kubernetes.io/projected/512340e2-4c40-48fd-ac35-8f84f1de5c6a-kube-api-access-5zcdm\") pod \"certified-operators-vkwz2\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.273295 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.274704 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.273948 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9w2bp"] Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.284454 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.311637 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.811571584 +0000 UTC m=+147.655003013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.337543 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9w2bp"] Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.343413 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.367761 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-catalog-content\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.380735 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-utilities\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.383439 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8hdg\" (UniqueName: \"kubernetes.io/projected/582c80c6-20e2-4a0e-9871-f1f4db232ca2-kube-api-access-l8hdg\") pod \"community-operators-5gkj5\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.397222 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.397879 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.897855993 +0000 UTC m=+147.741287432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.397913 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.397942 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-catalog-content\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.397968 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.397988 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9zn7\" (UniqueName: \"kubernetes.io/projected/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-kube-api-access-t9zn7\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.398012 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.398049 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-utilities\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.398071 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.402104 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:02.90208191 +0000 UTC m=+147.745513349 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.410839 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.413575 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.446137 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.448520 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:02 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:02 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:02 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.448572 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.486450 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.495940 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.501145 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.501558 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-catalog-content\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.501594 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9zn7\" (UniqueName: \"kubernetes.io/projected/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-kube-api-access-t9zn7\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.501644 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-utilities\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.502146 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-utilities\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.502190 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.002157766 +0000 UTC m=+147.845589205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.502448 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-catalog-content\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.551543 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9zn7\" (UniqueName: \"kubernetes.io/projected/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-kube-api-access-t9zn7\") pod \"certified-operators-9w2bp\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.605363 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.605833 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.105809695 +0000 UTC m=+147.949241134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.644905 4678 patch_prober.go:28] interesting pod/apiserver-76f77b778f-ck2d4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 06 10:39:02 crc kubenswrapper[4678]: [+]log ok Dec 06 10:39:02 crc kubenswrapper[4678]: [+]etcd ok Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 06 10:39:02 crc kubenswrapper[4678]: [-]poststarthook/generic-apiserver-start-informers failed: reason withheld Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/max-in-flight-filter ok Dec 06 10:39:02 crc kubenswrapper[4678]: [-]poststarthook/storage-object-count-tracker-hook failed: reason withheld Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 06 10:39:02 crc kubenswrapper[4678]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 06 10:39:02 crc kubenswrapper[4678]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/project.openshift.io-projectcache ok Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/openshift.io-startinformers ok Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 06 10:39:02 crc kubenswrapper[4678]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 06 10:39:02 crc kubenswrapper[4678]: livez check failed Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.644967 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" podUID="e1cd25ff-72f0-414f-a4e6-51c0d5eb5662" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.692892 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.695931 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.696106 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.708186 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.708619 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.208598706 +0000 UTC m=+148.052030145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.781186 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.809411 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.810058 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.310023514 +0000 UTC m=+148.153454953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.910755 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.410718012 +0000 UTC m=+148.254149451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.910827 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:02 crc kubenswrapper[4678]: I1206 10:39:02.911231 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:02 crc kubenswrapper[4678]: E1206 10:39:02.911625 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.411611932 +0000 UTC m=+148.255043371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.012198 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.012697 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.512672025 +0000 UTC m=+148.356103474 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.115213 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.115662 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.615625907 +0000 UTC m=+148.459057346 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.222582 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.226888 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.726823709 +0000 UTC m=+148.570255148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.227072 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.227658 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.727647204 +0000 UTC m=+148.571078643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.243629 4678 patch_prober.go:28] interesting pod/console-operator-58897d9998-lff5h container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.243714 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-lff5h" podUID="1438bcb1-f3fb-4979-91ba-2bce2952edb9" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.13:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.305690 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s8p98"] Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.327943 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.328213 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.828193252 +0000 UTC m=+148.671624691 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.328758 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" event={"ID":"3439565e-e37c-4962-a269-ddb8a48b3f43","Type":"ContainerStarted","Data":"728d0746709cb047c5245a19ab67a9a7602804a3c02a9178e3bb2ca9df3e52aa"} Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.341224 4678 generic.go:334] "Generic (PLEG): container finished" podID="4777f215-78ef-47d4-be38-0e2a750168fe" containerID="1ad63fcf3ea70355e95429e928b0d8ad6bb8651a83bdf2cf72bf9cc82a179b61" exitCode=0 Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.341329 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" event={"ID":"4777f215-78ef-47d4-be38-0e2a750168fe","Type":"ContainerDied","Data":"1ad63fcf3ea70355e95429e928b0d8ad6bb8651a83bdf2cf72bf9cc82a179b61"} Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.358890 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" event={"ID":"a0c46190-8e1a-4068-ba9f-9d3751787308","Type":"ContainerStarted","Data":"71b25b71f8fafd37448e9220a2ae95b5349fc6ee8ae53d52d914fc4ce3454043"} Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.377940 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-4scxr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.377994 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-4scxr" podUID="08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.378231 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-4scxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.378247 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4scxr" podUID="08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.387835 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rth98" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.395730 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.406931 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:03 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:03 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:03 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.407003 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.423645 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" podStartSLOduration=129.423622273 podStartE2EDuration="2m9.423622273s" podCreationTimestamp="2025-12-06 10:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:03.407370899 +0000 UTC m=+148.250802358" watchObservedRunningTime="2025-12-06 10:39:03.423622273 +0000 UTC m=+148.267053712" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.435016 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.449306 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:03.949281956 +0000 UTC m=+148.792713585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.466777 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-lff5h" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.542796 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.545410 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.045386422 +0000 UTC m=+148.888817861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.648111 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.648446 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.14843268 +0000 UTC m=+148.991864119 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.749586 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.749904 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.24988342 +0000 UTC m=+149.093314849 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.851715 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.852204 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.352187157 +0000 UTC m=+149.195618606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.923758 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vkwz2"] Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.953468 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.953958 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.453923556 +0000 UTC m=+149.297354995 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.959958 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:03 crc kubenswrapper[4678]: E1206 10:39:03.960609 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.460591349 +0000 UTC m=+149.304022788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.968671 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:39:03 crc kubenswrapper[4678]: I1206 10:39:03.969057 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.029055 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rfk99"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.030210 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.050172 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.074086 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.074644 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.574617703 +0000 UTC m=+149.418049142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.087443 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfk99"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.145702 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.166728 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.166771 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.169109 4678 patch_prober.go:28] interesting pod/console-f9d7485db-ll9jl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.169148 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ll9jl" podUID="f13f288c-c8dc-4df8-b89b-c325d088b30b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.194057 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.194149 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-catalog-content\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.194185 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-utilities\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.194222 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llgfk\" (UniqueName: \"kubernetes.io/projected/ebcefb39-28c6-4619-88c5-15537f549d0d-kube-api-access-llgfk\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.197481 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.697460075 +0000 UTC m=+149.540891514 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.303216 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.303496 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-catalog-content\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.303545 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-utilities\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.303572 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llgfk\" (UniqueName: \"kubernetes.io/projected/ebcefb39-28c6-4619-88c5-15537f549d0d-kube-api-access-llgfk\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.304137 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-catalog-content\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.304257 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.804232136 +0000 UTC m=+149.647663575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.304272 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-utilities\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.345694 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llgfk\" (UniqueName: \"kubernetes.io/projected/ebcefb39-28c6-4619-88c5-15537f549d0d-kube-api-access-llgfk\") pod \"redhat-marketplace-rfk99\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.364885 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.373108 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwz2" event={"ID":"512340e2-4c40-48fd-ac35-8f84f1de5c6a","Type":"ContainerStarted","Data":"b180765307011ca9426246c6433987fb42d0cbd6e4cb63957d200453c08a6b9e"} Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.394321 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"385668f8dd642fb84a81a2aacbbd5a23d89473ca6295295f78b7e681e2fa6c3c"} Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.404691 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.405121 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:04.905105246 +0000 UTC m=+149.748536685 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.412705 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:04 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:04 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:04 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.412774 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.422766 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" event={"ID":"a0c46190-8e1a-4068-ba9f-9d3751787308","Type":"ContainerStarted","Data":"cf78296eb7b3015e55ca2e306d6facab328746be8c2136708f1e8868d2fc8358"} Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.436763 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.440520 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nd8l6"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.441769 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.455808 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"cdca0c11942ef344d4b6d2cc0016ebdfd8fa2ae882964c57516b35ed71115e81"} Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.482475 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerStarted","Data":"ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609"} Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.485347 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerStarted","Data":"53e89eb026416fae766f04161feb1875f7d4ad1bdbad3c3e74ced6b319309362"} Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.486935 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nd8l6"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.503507 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2wz6s" podStartSLOduration=14.503488538 podStartE2EDuration="14.503488538s" podCreationTimestamp="2025-12-06 10:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:04.498113973 +0000 UTC m=+149.341545402" watchObservedRunningTime="2025-12-06 10:39:04.503488538 +0000 UTC m=+149.346919977" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.506326 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.506820 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.006797062 +0000 UTC m=+149.850228501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.510368 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5gkj5"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.529563 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gvw5z" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.531522 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.590236 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9w2bp"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.609399 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-utilities\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.609468 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.609524 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-catalog-content\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.609667 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm5wr\" (UniqueName: \"kubernetes.io/projected/9cd280dd-abfa-433b-bbbd-7f96542d01fc-kube-api-access-dm5wr\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.612216 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.11220071 +0000 UTC m=+149.955632149 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.710465 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.711160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-utilities\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.711211 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-catalog-content\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.711279 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm5wr\" (UniqueName: \"kubernetes.io/projected/9cd280dd-abfa-433b-bbbd-7f96542d01fc-kube-api-access-dm5wr\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.712497 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.212465659 +0000 UTC m=+150.055897098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.714579 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-utilities\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.715357 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-catalog-content\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.812739 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.813175 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.313156117 +0000 UTC m=+150.156587566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.855972 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm5wr\" (UniqueName: \"kubernetes.io/projected/9cd280dd-abfa-433b-bbbd-7f96542d01fc-kube-api-access-dm5wr\") pod \"redhat-marketplace-nd8l6\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.869563 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pxr6m"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.880852 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.884475 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.885204 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxr6m"] Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.914400 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:04 crc kubenswrapper[4678]: E1206 10:39:04.914904 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.414886006 +0000 UTC m=+150.258317435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:04 crc kubenswrapper[4678]: I1206 10:39:04.978554 4678 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.024370 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-utilities\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.024432 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.024469 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-catalog-content\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.024544 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffqkr\" (UniqueName: \"kubernetes.io/projected/3e103cce-344c-4158-8661-6f695eb8fcca-kube-api-access-ffqkr\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.024888 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.524873824 +0000 UTC m=+150.368305263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.045696 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-czg22"] Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.046794 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.070702 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czg22"] Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.114239 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.127745 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.128081 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-catalog-content\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.128164 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffqkr\" (UniqueName: \"kubernetes.io/projected/3e103cce-344c-4158-8661-6f695eb8fcca-kube-api-access-ffqkr\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.128199 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-utilities\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.128733 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-utilities\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.128988 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-catalog-content\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.129029 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.629010567 +0000 UTC m=+150.472442006 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.175695 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffqkr\" (UniqueName: \"kubernetes.io/projected/3e103cce-344c-4158-8661-6f695eb8fcca-kube-api-access-ffqkr\") pod \"redhat-operators-pxr6m\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.193145 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.193878 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.196865 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.200908 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.201101 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.221616 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.235573 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-utilities\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.235630 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8znz9\" (UniqueName: \"kubernetes.io/projected/2b942ed0-624a-47e7-8801-e495760b1eda-kube-api-access-8znz9\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.235710 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.235731 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-catalog-content\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.236121 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.736106 +0000 UTC m=+150.579537439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.344199 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.344679 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.84463779 +0000 UTC m=+150.688069239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.344919 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.344996 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.345026 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-catalog-content\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.345052 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.345123 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-utilities\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.345146 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8znz9\" (UniqueName: \"kubernetes.io/projected/2b942ed0-624a-47e7-8801-e495760b1eda-kube-api-access-8znz9\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.345890 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.845877365 +0000 UTC m=+150.689308804 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.347047 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-utilities\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.356547 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-catalog-content\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.388361 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8znz9\" (UniqueName: \"kubernetes.io/projected/2b942ed0-624a-47e7-8801-e495760b1eda-kube-api-access-8znz9\") pod \"redhat-operators-czg22\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.413167 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:05 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:05 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:05 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.413231 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.413846 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.451541 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.451815 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.451879 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.452405 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:05.952382997 +0000 UTC m=+150.795814436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.452443 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.470726 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.486801 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.557704 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.558202 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-06 10:39:06.058187683 +0000 UTC m=+150.901619112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-98427" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.559122 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7206ffd3a4d7266333ea393dc5f9d66b982ab6375d2a85a1041d947ddb0dacec"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.559197 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ecf0b3b02ab378a38d6e25ea6a77fce2ce2cd3403da1602e4f6147597eca7f39"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.559874 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.586252 4678 generic.go:334] "Generic (PLEG): container finished" podID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerID="b964b66a5b4dd52274c86e257a9c768a69c0f6d509107a01ac2d92e34a868769" exitCode=0 Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.586905 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9w2bp" event={"ID":"511df94f-24d2-43e7-a273-d0e2f2b5e6bc","Type":"ContainerDied","Data":"b964b66a5b4dd52274c86e257a9c768a69c0f6d509107a01ac2d92e34a868769"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.586939 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9w2bp" event={"ID":"511df94f-24d2-43e7-a273-d0e2f2b5e6bc","Type":"ContainerStarted","Data":"7d2f14ce76cd58dca88e640266e4b39b9311edf7fb73badcb0cb0f246ac09653"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.656506 4678 generic.go:334] "Generic (PLEG): container finished" podID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerID="9142876d5fe2c4f56a1b2a7e077ec670892a4ffb41bc96e7892fd1bc20670025" exitCode=0 Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.656628 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gkj5" event={"ID":"582c80c6-20e2-4a0e-9871-f1f4db232ca2","Type":"ContainerDied","Data":"9142876d5fe2c4f56a1b2a7e077ec670892a4ffb41bc96e7892fd1bc20670025"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.656707 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gkj5" event={"ID":"582c80c6-20e2-4a0e-9871-f1f4db232ca2","Type":"ContainerStarted","Data":"5241adc331e90f1036f2c7ab5fb478cb24fb4b8a1e94eeb596054c21560c4466"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.656765 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.659823 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4777f215-78ef-47d4-be38-0e2a750168fe-config-volume\") pod \"4777f215-78ef-47d4-be38-0e2a750168fe\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.659869 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4777f215-78ef-47d4-be38-0e2a750168fe-secret-volume\") pod \"4777f215-78ef-47d4-be38-0e2a750168fe\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.659896 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dpmn\" (UniqueName: \"kubernetes.io/projected/4777f215-78ef-47d4-be38-0e2a750168fe-kube-api-access-5dpmn\") pod \"4777f215-78ef-47d4-be38-0e2a750168fe\" (UID: \"4777f215-78ef-47d4-be38-0e2a750168fe\") " Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.660004 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:05 crc kubenswrapper[4678]: E1206 10:39:05.661733 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-06 10:39:06.161711243 +0000 UTC m=+151.005142682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.662927 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4777f215-78ef-47d4-be38-0e2a750168fe-config-volume" (OuterVolumeSpecName: "config-volume") pod "4777f215-78ef-47d4-be38-0e2a750168fe" (UID: "4777f215-78ef-47d4-be38-0e2a750168fe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.668169 4678 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-06T10:39:04.978605903Z","Handler":null,"Name":""} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.690329 4678 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.690369 4678 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.690646 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4777f215-78ef-47d4-be38-0e2a750168fe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4777f215-78ef-47d4-be38-0e2a750168fe" (UID: "4777f215-78ef-47d4-be38-0e2a750168fe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.694422 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4777f215-78ef-47d4-be38-0e2a750168fe-kube-api-access-5dpmn" (OuterVolumeSpecName: "kube-api-access-5dpmn") pod "4777f215-78ef-47d4-be38-0e2a750168fe" (UID: "4777f215-78ef-47d4-be38-0e2a750168fe"). InnerVolumeSpecName "kube-api-access-5dpmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.707783 4678 generic.go:334] "Generic (PLEG): container finished" podID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerID="ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609" exitCode=0 Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.707876 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerDied","Data":"ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.740958 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" event={"ID":"4777f215-78ef-47d4-be38-0e2a750168fe","Type":"ContainerDied","Data":"9ce49b55526ec52286791f13b14dadc3c172a85563dfa4471ee094a9072b7ef2"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.741036 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ce49b55526ec52286791f13b14dadc3c172a85563dfa4471ee094a9072b7ef2" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.741278 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.760743 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.760829 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4777f215-78ef-47d4-be38-0e2a750168fe-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.760843 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4777f215-78ef-47d4-be38-0e2a750168fe-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.760853 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dpmn\" (UniqueName: \"kubernetes.io/projected/4777f215-78ef-47d4-be38-0e2a750168fe-kube-api-access-5dpmn\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.799854 4678 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.799919 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.809385 4678 generic.go:334] "Generic (PLEG): container finished" podID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerID="24ccfacc1302566a11ed8aecaed79aef646ac8adee042aa56161d25c4b25adee" exitCode=0 Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.809551 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwz2" event={"ID":"512340e2-4c40-48fd-ac35-8f84f1de5c6a","Type":"ContainerDied","Data":"24ccfacc1302566a11ed8aecaed79aef646ac8adee042aa56161d25c4b25adee"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.850896 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ec942308df844e51e373debe3ae5ff376ad510c50af867234773abc2a64206c0"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.866647 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"4f7a667276500134a434f291eb1092c889a573325a865be49cce1baf944b33b1"} Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.919781 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfk99"] Dec 06 10:39:05 crc kubenswrapper[4678]: I1206 10:39:05.964953 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nd8l6"] Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.052318 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-98427\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.070057 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.127780 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.141314 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pxr6m"] Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.201812 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-czg22"] Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.286155 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.384181 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.407627 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:06 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:06 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:06 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.407684 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.737235 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-98427"] Dec 06 10:39:06 crc kubenswrapper[4678]: E1206 10:39:06.842211 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3e103cce_344c_4158_8661_6f695eb8fcca.slice/crio-91e7b7e6997c23038eaccb1d51b10a2b5a175d1882ee9efbb08deaa11ed8923a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b942ed0_624a_47e7_8801_e495760b1eda.slice/crio-7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2b942ed0_624a_47e7_8801_e495760b1eda.slice/crio-conmon-7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3.scope\": RecentStats: unable to find data in memory cache]" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.842457 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 10:39:06 crc kubenswrapper[4678]: E1206 10:39:06.842767 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4777f215-78ef-47d4-be38-0e2a750168fe" containerName="collect-profiles" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.842786 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4777f215-78ef-47d4-be38-0e2a750168fe" containerName="collect-profiles" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.842896 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4777f215-78ef-47d4-be38-0e2a750168fe" containerName="collect-profiles" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.843319 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.848141 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.850777 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.852721 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.861870 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.901028 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-ck2d4" Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.937063 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b968f7f6-a0df-409c-b0fa-9152c3d23e73","Type":"ContainerStarted","Data":"ac7027ffe6970919334bc65dc3adde10066dc6727405de98e40bb71424dc91af"} Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.957679 4678 generic.go:334] "Generic (PLEG): container finished" podID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerID="c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea" exitCode=0 Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.957765 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nd8l6" event={"ID":"9cd280dd-abfa-433b-bbbd-7f96542d01fc","Type":"ContainerDied","Data":"c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea"} Dec 06 10:39:06 crc kubenswrapper[4678]: I1206 10:39:06.957808 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nd8l6" event={"ID":"9cd280dd-abfa-433b-bbbd-7f96542d01fc","Type":"ContainerStarted","Data":"c21b7cac4edcb57b1f97305df2432596675d8763ef2064289e154b1dfec3e060"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.015490 4678 generic.go:334] "Generic (PLEG): container finished" podID="2b942ed0-624a-47e7-8801-e495760b1eda" containerID="7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3" exitCode=0 Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.016289 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czg22" event={"ID":"2b942ed0-624a-47e7-8801-e495760b1eda","Type":"ContainerDied","Data":"7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.016321 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czg22" event={"ID":"2b942ed0-624a-47e7-8801-e495760b1eda","Type":"ContainerStarted","Data":"9c38062abf21476e31316b215b90a658c76367f69704c53900ae53820f756506"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.019095 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/200cb88d-b8ed-4753-9f5b-90381b983561-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.019118 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/200cb88d-b8ed-4753-9f5b-90381b983561-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.048131 4678 generic.go:334] "Generic (PLEG): container finished" podID="3e103cce-344c-4158-8661-6f695eb8fcca" containerID="91e7b7e6997c23038eaccb1d51b10a2b5a175d1882ee9efbb08deaa11ed8923a" exitCode=0 Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.048240 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxr6m" event={"ID":"3e103cce-344c-4158-8661-6f695eb8fcca","Type":"ContainerDied","Data":"91e7b7e6997c23038eaccb1d51b10a2b5a175d1882ee9efbb08deaa11ed8923a"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.048272 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxr6m" event={"ID":"3e103cce-344c-4158-8661-6f695eb8fcca","Type":"ContainerStarted","Data":"9ab40a73b16e556d35089072fcc79038d09cbf549afd9e7360bc9b07bf18b062"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.072752 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-98427" event={"ID":"140fee33-c778-4f79-941e-430f5ae44bd9","Type":"ContainerStarted","Data":"ea1b9c05bc813fc2bbc1e0fa541b82c55066e8918caf3d7f29b3b2e57456f97f"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.092897 4678 generic.go:334] "Generic (PLEG): container finished" podID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerID="dac94fb835d764a87d50534f5acf212c5f9ad4b4a6ab704d1724e84974387bdf" exitCode=0 Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.103090 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfk99" event={"ID":"ebcefb39-28c6-4619-88c5-15537f549d0d","Type":"ContainerDied","Data":"dac94fb835d764a87d50534f5acf212c5f9ad4b4a6ab704d1724e84974387bdf"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.103146 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfk99" event={"ID":"ebcefb39-28c6-4619-88c5-15537f549d0d","Type":"ContainerStarted","Data":"c24ec533b4bd06f94c895d60d8215f696dbc16328053d3a690ef6f7075aa57be"} Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.121321 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/200cb88d-b8ed-4753-9f5b-90381b983561-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.121360 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/200cb88d-b8ed-4753-9f5b-90381b983561-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.124081 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/200cb88d-b8ed-4753-9f5b-90381b983561-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.189474 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/200cb88d-b8ed-4753-9f5b-90381b983561-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.202240 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.414222 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:07 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:07 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:07 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.414640 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.510159 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 06 10:39:07 crc kubenswrapper[4678]: I1206 10:39:07.758514 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.123593 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"200cb88d-b8ed-4753-9f5b-90381b983561","Type":"ContainerStarted","Data":"2d1352274912a16813ddd678cf56fe4aa4734aaa36dcf6c71706a00a07ffff85"} Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.150910 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-98427" event={"ID":"140fee33-c778-4f79-941e-430f5ae44bd9","Type":"ContainerStarted","Data":"5cb32ca746c5a658695264ebee9b62b3fbf74f9ca8b4be97e79bf1aebeb46e11"} Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.151069 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.162035 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b968f7f6-a0df-409c-b0fa-9152c3d23e73","Type":"ContainerStarted","Data":"f0e67c14b1fad5dcd0bf7a9b93d764bb7b112678fdb3ee28702d68e112df21f4"} Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.187582 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-98427" podStartSLOduration=135.187560024 podStartE2EDuration="2m15.187560024s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:08.180202454 +0000 UTC m=+153.023633893" watchObservedRunningTime="2025-12-06 10:39:08.187560024 +0000 UTC m=+153.030991463" Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.207904 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.207875843 podStartE2EDuration="3.207875843s" podCreationTimestamp="2025-12-06 10:39:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:08.203338915 +0000 UTC m=+153.046770354" watchObservedRunningTime="2025-12-06 10:39:08.207875843 +0000 UTC m=+153.051307282" Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.401261 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:08 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:08 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:08 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:08 crc kubenswrapper[4678]: I1206 10:39:08.401343 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:09 crc kubenswrapper[4678]: I1206 10:39:09.198082 4678 generic.go:334] "Generic (PLEG): container finished" podID="b968f7f6-a0df-409c-b0fa-9152c3d23e73" containerID="f0e67c14b1fad5dcd0bf7a9b93d764bb7b112678fdb3ee28702d68e112df21f4" exitCode=0 Dec 06 10:39:09 crc kubenswrapper[4678]: I1206 10:39:09.199584 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b968f7f6-a0df-409c-b0fa-9152c3d23e73","Type":"ContainerDied","Data":"f0e67c14b1fad5dcd0bf7a9b93d764bb7b112678fdb3ee28702d68e112df21f4"} Dec 06 10:39:09 crc kubenswrapper[4678]: I1206 10:39:09.220336 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"200cb88d-b8ed-4753-9f5b-90381b983561","Type":"ContainerStarted","Data":"c998db890322079121e7040ed5c1a4ad3ddce5b4b9a420bf46550818621014d2"} Dec 06 10:39:09 crc kubenswrapper[4678]: I1206 10:39:09.400036 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:09 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:09 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:09 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:09 crc kubenswrapper[4678]: I1206 10:39:09.400109 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.254165 4678 generic.go:334] "Generic (PLEG): container finished" podID="200cb88d-b8ed-4753-9f5b-90381b983561" containerID="c998db890322079121e7040ed5c1a4ad3ddce5b4b9a420bf46550818621014d2" exitCode=0 Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.254826 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"200cb88d-b8ed-4753-9f5b-90381b983561","Type":"ContainerDied","Data":"c998db890322079121e7040ed5c1a4ad3ddce5b4b9a420bf46550818621014d2"} Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.398990 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:10 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:10 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:10 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.399072 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.647692 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.814381 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kubelet-dir\") pod \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.814486 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kube-api-access\") pod \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\" (UID: \"b968f7f6-a0df-409c-b0fa-9152c3d23e73\") " Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.815166 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b968f7f6-a0df-409c-b0fa-9152c3d23e73" (UID: "b968f7f6-a0df-409c-b0fa-9152c3d23e73"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.835357 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b968f7f6-a0df-409c-b0fa-9152c3d23e73" (UID: "b968f7f6-a0df-409c-b0fa-9152c3d23e73"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.915983 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:10 crc kubenswrapper[4678]: I1206 10:39:10.916020 4678 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b968f7f6-a0df-409c-b0fa-9152c3d23e73-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.317705 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.320830 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b968f7f6-a0df-409c-b0fa-9152c3d23e73","Type":"ContainerDied","Data":"ac7027ffe6970919334bc65dc3adde10066dc6727405de98e40bb71424dc91af"} Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.320884 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac7027ffe6970919334bc65dc3adde10066dc6727405de98e40bb71424dc91af" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.404668 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:11 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:11 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:11 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.404753 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.644075 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.660539 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/200cb88d-b8ed-4753-9f5b-90381b983561-kubelet-dir\") pod \"200cb88d-b8ed-4753-9f5b-90381b983561\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.660643 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/200cb88d-b8ed-4753-9f5b-90381b983561-kube-api-access\") pod \"200cb88d-b8ed-4753-9f5b-90381b983561\" (UID: \"200cb88d-b8ed-4753-9f5b-90381b983561\") " Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.661882 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/200cb88d-b8ed-4753-9f5b-90381b983561-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "200cb88d-b8ed-4753-9f5b-90381b983561" (UID: "200cb88d-b8ed-4753-9f5b-90381b983561"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.681173 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/200cb88d-b8ed-4753-9f5b-90381b983561-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "200cb88d-b8ed-4753-9f5b-90381b983561" (UID: "200cb88d-b8ed-4753-9f5b-90381b983561"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.762039 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/200cb88d-b8ed-4753-9f5b-90381b983561-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.762074 4678 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/200cb88d-b8ed-4753-9f5b-90381b983561-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:11 crc kubenswrapper[4678]: I1206 10:39:11.875273 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-m42c5" Dec 06 10:39:12 crc kubenswrapper[4678]: I1206 10:39:12.356930 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"200cb88d-b8ed-4753-9f5b-90381b983561","Type":"ContainerDied","Data":"2d1352274912a16813ddd678cf56fe4aa4734aaa36dcf6c71706a00a07ffff85"} Dec 06 10:39:12 crc kubenswrapper[4678]: I1206 10:39:12.356982 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d1352274912a16813ddd678cf56fe4aa4734aaa36dcf6c71706a00a07ffff85" Dec 06 10:39:12 crc kubenswrapper[4678]: I1206 10:39:12.356983 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 06 10:39:12 crc kubenswrapper[4678]: I1206 10:39:12.399392 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:12 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:12 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:12 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:12 crc kubenswrapper[4678]: I1206 10:39:12.399459 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:13 crc kubenswrapper[4678]: I1206 10:39:13.377740 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-4scxr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 10:39:13 crc kubenswrapper[4678]: I1206 10:39:13.377820 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-4scxr" podUID="08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 10:39:13 crc kubenswrapper[4678]: I1206 10:39:13.378604 4678 patch_prober.go:28] interesting pod/downloads-7954f5f757-4scxr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 06 10:39:13 crc kubenswrapper[4678]: I1206 10:39:13.378626 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-4scxr" podUID="08d3b6b1-f2d0-4cd6-b6e7-f6410738bfa4" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 06 10:39:13 crc kubenswrapper[4678]: I1206 10:39:13.399218 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:13 crc kubenswrapper[4678]: [-]has-synced failed: reason withheld Dec 06 10:39:13 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:13 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:13 crc kubenswrapper[4678]: I1206 10:39:13.399785 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:14 crc kubenswrapper[4678]: I1206 10:39:14.161130 4678 patch_prober.go:28] interesting pod/console-f9d7485db-ll9jl container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Dec 06 10:39:14 crc kubenswrapper[4678]: I1206 10:39:14.161210 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ll9jl" podUID="f13f288c-c8dc-4df8-b89b-c325d088b30b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Dec 06 10:39:14 crc kubenswrapper[4678]: I1206 10:39:14.398195 4678 patch_prober.go:28] interesting pod/router-default-5444994796-z4j5s container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 06 10:39:14 crc kubenswrapper[4678]: [+]has-synced ok Dec 06 10:39:14 crc kubenswrapper[4678]: [+]process-running ok Dec 06 10:39:14 crc kubenswrapper[4678]: healthz check failed Dec 06 10:39:14 crc kubenswrapper[4678]: I1206 10:39:14.398278 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-z4j5s" podUID="40b9fa22-f2c6-4fb6-993a-65b61c12fed7" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:39:15 crc kubenswrapper[4678]: I1206 10:39:15.400613 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:39:15 crc kubenswrapper[4678]: I1206 10:39:15.404594 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-z4j5s" Dec 06 10:39:16 crc kubenswrapper[4678]: I1206 10:39:16.788119 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:39:16 crc kubenswrapper[4678]: I1206 10:39:16.814379 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5fdea9d8-255a-4ec7-8095-4489828f4136-metrics-certs\") pod \"network-metrics-daemon-wfvj4\" (UID: \"5fdea9d8-255a-4ec7-8095-4489828f4136\") " pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:39:16 crc kubenswrapper[4678]: I1206 10:39:16.899114 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-wfvj4" Dec 06 10:39:23 crc kubenswrapper[4678]: I1206 10:39:23.381601 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-4scxr" Dec 06 10:39:24 crc kubenswrapper[4678]: I1206 10:39:24.165756 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:39:24 crc kubenswrapper[4678]: I1206 10:39:24.171591 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:39:26 crc kubenswrapper[4678]: I1206 10:39:26.296015 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:39:29 crc kubenswrapper[4678]: I1206 10:39:29.505301 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:39:29 crc kubenswrapper[4678]: I1206 10:39:29.505857 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:39:33 crc kubenswrapper[4678]: I1206 10:39:33.805477 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pl9h6" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.243335 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 10:39:40 crc kubenswrapper[4678]: E1206 10:39:40.246431 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="200cb88d-b8ed-4753-9f5b-90381b983561" containerName="pruner" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.246445 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="200cb88d-b8ed-4753-9f5b-90381b983561" containerName="pruner" Dec 06 10:39:40 crc kubenswrapper[4678]: E1206 10:39:40.246461 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b968f7f6-a0df-409c-b0fa-9152c3d23e73" containerName="pruner" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.246469 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b968f7f6-a0df-409c-b0fa-9152c3d23e73" containerName="pruner" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.246635 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b968f7f6-a0df-409c-b0fa-9152c3d23e73" containerName="pruner" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.246645 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="200cb88d-b8ed-4753-9f5b-90381b983561" containerName="pruner" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.247054 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.247520 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.250366 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.250594 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.357094 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa9995fb-c5e9-40b6-9415-14a32349cf94-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.357552 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa9995fb-c5e9-40b6-9415-14a32349cf94-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.458653 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa9995fb-c5e9-40b6-9415-14a32349cf94-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.458726 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa9995fb-c5e9-40b6-9415-14a32349cf94-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.458747 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa9995fb-c5e9-40b6-9415-14a32349cf94-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.486766 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa9995fb-c5e9-40b6-9415-14a32349cf94-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:40 crc kubenswrapper[4678]: I1206 10:39:40.571008 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:42 crc kubenswrapper[4678]: I1206 10:39:42.700898 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 06 10:39:43 crc kubenswrapper[4678]: E1206 10:39:43.242070 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 06 10:39:43 crc kubenswrapper[4678]: E1206 10:39:43.242919 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8znz9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-czg22_openshift-marketplace(2b942ed0-624a-47e7-8801-e495760b1eda): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 10:39:43 crc kubenswrapper[4678]: E1206 10:39:43.244182 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-czg22" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" Dec 06 10:39:43 crc kubenswrapper[4678]: E1206 10:39:43.258762 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 06 10:39:43 crc kubenswrapper[4678]: E1206 10:39:43.258945 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r66k5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-s8p98_openshift-marketplace(29fca54e-1ec1-484b-b3c4-431ecaedd87b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 10:39:43 crc kubenswrapper[4678]: E1206 10:39:43.260149 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-s8p98" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.531658 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-czg22" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.544072 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.544254 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-llgfk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-rfk99_openshift-marketplace(ebcefb39-28c6-4619-88c5-15537f549d0d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.544654 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-s8p98" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.546221 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-rfk99" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.694220 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.695022 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dm5wr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nd8l6_openshift-marketplace(9cd280dd-abfa-433b-bbbd-7f96542d01fc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 10:39:44 crc kubenswrapper[4678]: E1206 10:39:44.696223 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nd8l6" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.021146 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.024102 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.033285 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kube-api-access\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.033346 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.033407 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-var-lock\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.062173 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.134768 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-var-lock\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.134885 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kube-api-access\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.134906 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-var-lock\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.134945 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.134918 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.161748 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kube-api-access\") pod \"installer-9-crc\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: I1206 10:39:46.365284 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:39:46 crc kubenswrapper[4678]: E1206 10:39:46.971255 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-nd8l6" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" Dec 06 10:39:46 crc kubenswrapper[4678]: E1206 10:39:46.971313 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-rfk99" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.067442 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.072899 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zcdm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vkwz2_openshift-marketplace(512340e2-4c40-48fd-ac35-8f84f1de5c6a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.074057 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vkwz2" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.145710 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.145903 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t9zn7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9w2bp_openshift-marketplace(511df94f-24d2-43e7-a273-d0e2f2b5e6bc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.147394 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9w2bp" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" Dec 06 10:39:47 crc kubenswrapper[4678]: I1206 10:39:47.549159 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-wfvj4"] Dec 06 10:39:47 crc kubenswrapper[4678]: W1206 10:39:47.556932 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fdea9d8_255a_4ec7_8095_4489828f4136.slice/crio-9bf24bbf8ca92e88310a08f684ff606d81eeb34616b7379bbbcffd8be2d151f3 WatchSource:0}: Error finding container 9bf24bbf8ca92e88310a08f684ff606d81eeb34616b7379bbbcffd8be2d151f3: Status 404 returned error can't find the container with id 9bf24bbf8ca92e88310a08f684ff606d81eeb34616b7379bbbcffd8be2d151f3 Dec 06 10:39:47 crc kubenswrapper[4678]: I1206 10:39:47.645740 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 06 10:39:47 crc kubenswrapper[4678]: I1206 10:39:47.646097 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxr6m" event={"ID":"3e103cce-344c-4158-8661-6f695eb8fcca","Type":"ContainerStarted","Data":"998bb5ae4d5da43edc078c341ffcb4565cb2232fe2ec423fb09e3a9eae5326d9"} Dec 06 10:39:47 crc kubenswrapper[4678]: I1206 10:39:47.654423 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" event={"ID":"5fdea9d8-255a-4ec7-8095-4489828f4136","Type":"ContainerStarted","Data":"9bf24bbf8ca92e88310a08f684ff606d81eeb34616b7379bbbcffd8be2d151f3"} Dec 06 10:39:47 crc kubenswrapper[4678]: I1206 10:39:47.664764 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 06 10:39:47 crc kubenswrapper[4678]: I1206 10:39:47.672599 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gkj5" event={"ID":"582c80c6-20e2-4a0e-9871-f1f4db232ca2","Type":"ContainerStarted","Data":"fa4e40e63f9f620c299588aa86f937364405ea4b5a096e478a468a51bb3219bf"} Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.676779 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vkwz2" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" Dec 06 10:39:47 crc kubenswrapper[4678]: E1206 10:39:47.677239 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9w2bp" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" Dec 06 10:39:47 crc kubenswrapper[4678]: W1206 10:39:47.720157 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podaa9995fb_c5e9_40b6_9415_14a32349cf94.slice/crio-3d382dc2fea813b7af61431fccf04a6f1618689564c40a23652f46faf20eb74d WatchSource:0}: Error finding container 3d382dc2fea813b7af61431fccf04a6f1618689564c40a23652f46faf20eb74d: Status 404 returned error can't find the container with id 3d382dc2fea813b7af61431fccf04a6f1618689564c40a23652f46faf20eb74d Dec 06 10:39:48 crc kubenswrapper[4678]: I1206 10:39:48.680327 4678 generic.go:334] "Generic (PLEG): container finished" podID="3e103cce-344c-4158-8661-6f695eb8fcca" containerID="998bb5ae4d5da43edc078c341ffcb4565cb2232fe2ec423fb09e3a9eae5326d9" exitCode=0 Dec 06 10:39:48 crc kubenswrapper[4678]: I1206 10:39:48.680823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxr6m" event={"ID":"3e103cce-344c-4158-8661-6f695eb8fcca","Type":"ContainerDied","Data":"998bb5ae4d5da43edc078c341ffcb4565cb2232fe2ec423fb09e3a9eae5326d9"} Dec 06 10:39:48 crc kubenswrapper[4678]: I1206 10:39:48.682892 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aa9995fb-c5e9-40b6-9415-14a32349cf94","Type":"ContainerStarted","Data":"3d382dc2fea813b7af61431fccf04a6f1618689564c40a23652f46faf20eb74d"} Dec 06 10:39:48 crc kubenswrapper[4678]: I1206 10:39:48.684746 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50","Type":"ContainerStarted","Data":"d3e05b1d046788ef16d4d29ae35346169c75a31634cc4e2040adf01590bdb515"} Dec 06 10:39:48 crc kubenswrapper[4678]: I1206 10:39:48.686636 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" event={"ID":"5fdea9d8-255a-4ec7-8095-4489828f4136","Type":"ContainerStarted","Data":"ec69ea59db59f967699a2da4277e0402f61d46c3f7e1233fed318483de347ffc"} Dec 06 10:39:48 crc kubenswrapper[4678]: I1206 10:39:48.690618 4678 generic.go:334] "Generic (PLEG): container finished" podID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerID="fa4e40e63f9f620c299588aa86f937364405ea4b5a096e478a468a51bb3219bf" exitCode=0 Dec 06 10:39:48 crc kubenswrapper[4678]: I1206 10:39:48.690659 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gkj5" event={"ID":"582c80c6-20e2-4a0e-9871-f1f4db232ca2","Type":"ContainerDied","Data":"fa4e40e63f9f620c299588aa86f937364405ea4b5a096e478a468a51bb3219bf"} Dec 06 10:39:49 crc kubenswrapper[4678]: I1206 10:39:49.697744 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aa9995fb-c5e9-40b6-9415-14a32349cf94","Type":"ContainerStarted","Data":"067add553c9815d324f37b76d882628b6f7ecb0ba93bcf422cc7b4ff6f87a7d2"} Dec 06 10:39:49 crc kubenswrapper[4678]: I1206 10:39:49.705234 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50","Type":"ContainerStarted","Data":"35c971ec85e7a10afee376d8db22213cba410f3ce391bd58cc080791b688a87a"} Dec 06 10:39:49 crc kubenswrapper[4678]: I1206 10:39:49.718106 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=9.71808703 podStartE2EDuration="9.71808703s" podCreationTimestamp="2025-12-06 10:39:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:49.711476958 +0000 UTC m=+194.554908407" watchObservedRunningTime="2025-12-06 10:39:49.71808703 +0000 UTC m=+194.561518469" Dec 06 10:39:50 crc kubenswrapper[4678]: I1206 10:39:50.714163 4678 generic.go:334] "Generic (PLEG): container finished" podID="aa9995fb-c5e9-40b6-9415-14a32349cf94" containerID="067add553c9815d324f37b76d882628b6f7ecb0ba93bcf422cc7b4ff6f87a7d2" exitCode=0 Dec 06 10:39:50 crc kubenswrapper[4678]: I1206 10:39:50.714260 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aa9995fb-c5e9-40b6-9415-14a32349cf94","Type":"ContainerDied","Data":"067add553c9815d324f37b76d882628b6f7ecb0ba93bcf422cc7b4ff6f87a7d2"} Dec 06 10:39:50 crc kubenswrapper[4678]: I1206 10:39:50.717385 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-wfvj4" event={"ID":"5fdea9d8-255a-4ec7-8095-4489828f4136","Type":"ContainerStarted","Data":"001f3789ca637172a647df5b3291758188558975b94c1828e239bf397a0583af"} Dec 06 10:39:50 crc kubenswrapper[4678]: I1206 10:39:50.769425 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=4.769406756 podStartE2EDuration="4.769406756s" podCreationTimestamp="2025-12-06 10:39:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:50.751461559 +0000 UTC m=+195.594892998" watchObservedRunningTime="2025-12-06 10:39:50.769406756 +0000 UTC m=+195.612838195" Dec 06 10:39:50 crc kubenswrapper[4678]: I1206 10:39:50.769585 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-wfvj4" podStartSLOduration=177.769580972 podStartE2EDuration="2m57.769580972s" podCreationTimestamp="2025-12-06 10:36:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:39:50.766360333 +0000 UTC m=+195.609791772" watchObservedRunningTime="2025-12-06 10:39:50.769580972 +0000 UTC m=+195.613012411" Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.010638 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.041219 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa9995fb-c5e9-40b6-9415-14a32349cf94-kube-api-access\") pod \"aa9995fb-c5e9-40b6-9415-14a32349cf94\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.041296 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa9995fb-c5e9-40b6-9415-14a32349cf94-kubelet-dir\") pod \"aa9995fb-c5e9-40b6-9415-14a32349cf94\" (UID: \"aa9995fb-c5e9-40b6-9415-14a32349cf94\") " Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.041611 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa9995fb-c5e9-40b6-9415-14a32349cf94-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "aa9995fb-c5e9-40b6-9415-14a32349cf94" (UID: "aa9995fb-c5e9-40b6-9415-14a32349cf94"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.051778 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa9995fb-c5e9-40b6-9415-14a32349cf94-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "aa9995fb-c5e9-40b6-9415-14a32349cf94" (UID: "aa9995fb-c5e9-40b6-9415-14a32349cf94"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.142966 4678 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa9995fb-c5e9-40b6-9415-14a32349cf94-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.143052 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa9995fb-c5e9-40b6-9415-14a32349cf94-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.732440 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"aa9995fb-c5e9-40b6-9415-14a32349cf94","Type":"ContainerDied","Data":"3d382dc2fea813b7af61431fccf04a6f1618689564c40a23652f46faf20eb74d"} Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.732480 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d382dc2fea813b7af61431fccf04a6f1618689564c40a23652f46faf20eb74d" Dec 06 10:39:52 crc kubenswrapper[4678]: I1206 10:39:52.732551 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 06 10:39:53 crc kubenswrapper[4678]: I1206 10:39:53.740468 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gkj5" event={"ID":"582c80c6-20e2-4a0e-9871-f1f4db232ca2","Type":"ContainerStarted","Data":"fcf0aa3cfc5bf9076e78306cdcc1e3b7c2e21229cbb20655bcbd3f3c8f4236ce"} Dec 06 10:39:53 crc kubenswrapper[4678]: I1206 10:39:53.757956 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5gkj5" podStartSLOduration=6.374326912 podStartE2EDuration="52.757935527s" podCreationTimestamp="2025-12-06 10:39:01 +0000 UTC" firstStartedPulling="2025-12-06 10:39:05.669093564 +0000 UTC m=+150.512525003" lastFinishedPulling="2025-12-06 10:39:52.052702179 +0000 UTC m=+196.896133618" observedRunningTime="2025-12-06 10:39:53.756398862 +0000 UTC m=+198.599830301" watchObservedRunningTime="2025-12-06 10:39:53.757935527 +0000 UTC m=+198.601366966" Dec 06 10:39:55 crc kubenswrapper[4678]: I1206 10:39:55.752359 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxr6m" event={"ID":"3e103cce-344c-4158-8661-6f695eb8fcca","Type":"ContainerStarted","Data":"b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381"} Dec 06 10:39:57 crc kubenswrapper[4678]: I1206 10:39:57.501438 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pxr6m" podStartSLOduration=5.819065556 podStartE2EDuration="53.501397753s" podCreationTimestamp="2025-12-06 10:39:04 +0000 UTC" firstStartedPulling="2025-12-06 10:39:07.070391007 +0000 UTC m=+151.913822446" lastFinishedPulling="2025-12-06 10:39:54.752723204 +0000 UTC m=+199.596154643" observedRunningTime="2025-12-06 10:39:56.7727198 +0000 UTC m=+201.616151239" watchObservedRunningTime="2025-12-06 10:39:57.501397753 +0000 UTC m=+202.344829232" Dec 06 10:39:58 crc kubenswrapper[4678]: I1206 10:39:58.768097 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czg22" event={"ID":"2b942ed0-624a-47e7-8801-e495760b1eda","Type":"ContainerStarted","Data":"229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600"} Dec 06 10:39:59 crc kubenswrapper[4678]: I1206 10:39:59.505341 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:39:59 crc kubenswrapper[4678]: I1206 10:39:59.505439 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:39:59 crc kubenswrapper[4678]: I1206 10:39:59.505539 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:39:59 crc kubenswrapper[4678]: I1206 10:39:59.506309 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 10:39:59 crc kubenswrapper[4678]: I1206 10:39:59.506453 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55" gracePeriod=600 Dec 06 10:39:59 crc kubenswrapper[4678]: I1206 10:39:59.777078 4678 generic.go:334] "Generic (PLEG): container finished" podID="2b942ed0-624a-47e7-8801-e495760b1eda" containerID="229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600" exitCode=0 Dec 06 10:39:59 crc kubenswrapper[4678]: I1206 10:39:59.777160 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czg22" event={"ID":"2b942ed0-624a-47e7-8801-e495760b1eda","Type":"ContainerDied","Data":"229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600"} Dec 06 10:40:00 crc kubenswrapper[4678]: I1206 10:40:00.785768 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55" exitCode=0 Dec 06 10:40:00 crc kubenswrapper[4678]: I1206 10:40:00.786181 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55"} Dec 06 10:40:00 crc kubenswrapper[4678]: I1206 10:40:00.787814 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czg22" event={"ID":"2b942ed0-624a-47e7-8801-e495760b1eda","Type":"ContainerStarted","Data":"32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4"} Dec 06 10:40:00 crc kubenswrapper[4678]: I1206 10:40:00.809213 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-czg22" podStartSLOduration=3.203415284 podStartE2EDuration="56.809187577s" podCreationTimestamp="2025-12-06 10:39:04 +0000 UTC" firstStartedPulling="2025-12-06 10:39:07.028869487 +0000 UTC m=+151.872300926" lastFinishedPulling="2025-12-06 10:40:00.63464178 +0000 UTC m=+205.478073219" observedRunningTime="2025-12-06 10:40:00.807424724 +0000 UTC m=+205.650856163" watchObservedRunningTime="2025-12-06 10:40:00.809187577 +0000 UTC m=+205.652619016" Dec 06 10:40:01 crc kubenswrapper[4678]: I1206 10:40:01.796815 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9w2bp" event={"ID":"511df94f-24d2-43e7-a273-d0e2f2b5e6bc","Type":"ContainerStarted","Data":"75545da2a222326d142ac13ac481a4bf79ab37a9810668ceb154dbca84e0b18d"} Dec 06 10:40:01 crc kubenswrapper[4678]: I1206 10:40:01.798532 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerStarted","Data":"9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0"} Dec 06 10:40:01 crc kubenswrapper[4678]: I1206 10:40:01.800426 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"d59ce089befb53a3c54880fdb83e41fe8e063b3a198e12260fb6860590ff9b88"} Dec 06 10:40:01 crc kubenswrapper[4678]: I1206 10:40:01.805123 4678 generic.go:334] "Generic (PLEG): container finished" podID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerID="6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352" exitCode=0 Dec 06 10:40:01 crc kubenswrapper[4678]: I1206 10:40:01.805221 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nd8l6" event={"ID":"9cd280dd-abfa-433b-bbbd-7f96542d01fc","Type":"ContainerDied","Data":"6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352"} Dec 06 10:40:02 crc kubenswrapper[4678]: I1206 10:40:02.697561 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:40:02 crc kubenswrapper[4678]: I1206 10:40:02.698150 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:40:02 crc kubenswrapper[4678]: I1206 10:40:02.812894 4678 generic.go:334] "Generic (PLEG): container finished" podID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerID="75545da2a222326d142ac13ac481a4bf79ab37a9810668ceb154dbca84e0b18d" exitCode=0 Dec 06 10:40:02 crc kubenswrapper[4678]: I1206 10:40:02.812980 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9w2bp" event={"ID":"511df94f-24d2-43e7-a273-d0e2f2b5e6bc","Type":"ContainerDied","Data":"75545da2a222326d142ac13ac481a4bf79ab37a9810668ceb154dbca84e0b18d"} Dec 06 10:40:02 crc kubenswrapper[4678]: I1206 10:40:02.817008 4678 generic.go:334] "Generic (PLEG): container finished" podID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerID="9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0" exitCode=0 Dec 06 10:40:02 crc kubenswrapper[4678]: I1206 10:40:02.817720 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerDied","Data":"9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0"} Dec 06 10:40:03 crc kubenswrapper[4678]: I1206 10:40:03.003386 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:40:03 crc kubenswrapper[4678]: I1206 10:40:03.055173 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:40:03 crc kubenswrapper[4678]: I1206 10:40:03.827943 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9w2bp" event={"ID":"511df94f-24d2-43e7-a273-d0e2f2b5e6bc","Type":"ContainerStarted","Data":"a1306e68affc6040ffc13909a68ab91a75ed1ed6e6cf5895c71e20dfe3488f5d"} Dec 06 10:40:03 crc kubenswrapper[4678]: I1206 10:40:03.833463 4678 generic.go:334] "Generic (PLEG): container finished" podID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerID="2e381988ada7a179645da3cb3435104419646a6ddbd9e8a7fa50611da4a2ffaf" exitCode=0 Dec 06 10:40:03 crc kubenswrapper[4678]: I1206 10:40:03.833519 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfk99" event={"ID":"ebcefb39-28c6-4619-88c5-15537f549d0d","Type":"ContainerDied","Data":"2e381988ada7a179645da3cb3435104419646a6ddbd9e8a7fa50611da4a2ffaf"} Dec 06 10:40:03 crc kubenswrapper[4678]: I1206 10:40:03.871006 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9w2bp" podStartSLOduration=4.282380802 podStartE2EDuration="1m1.870974872s" podCreationTimestamp="2025-12-06 10:39:02 +0000 UTC" firstStartedPulling="2025-12-06 10:39:05.600093588 +0000 UTC m=+150.443525027" lastFinishedPulling="2025-12-06 10:40:03.188687658 +0000 UTC m=+208.032119097" observedRunningTime="2025-12-06 10:40:03.866045397 +0000 UTC m=+208.709476836" watchObservedRunningTime="2025-12-06 10:40:03.870974872 +0000 UTC m=+208.714406311" Dec 06 10:40:04 crc kubenswrapper[4678]: I1206 10:40:04.842409 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nd8l6" event={"ID":"9cd280dd-abfa-433b-bbbd-7f96542d01fc","Type":"ContainerStarted","Data":"13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a"} Dec 06 10:40:04 crc kubenswrapper[4678]: I1206 10:40:04.846068 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerStarted","Data":"c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911"} Dec 06 10:40:04 crc kubenswrapper[4678]: I1206 10:40:04.870833 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nd8l6" podStartSLOduration=4.587054727 podStartE2EDuration="1m0.870809949s" podCreationTimestamp="2025-12-06 10:39:04 +0000 UTC" firstStartedPulling="2025-12-06 10:39:06.971883837 +0000 UTC m=+151.815315296" lastFinishedPulling="2025-12-06 10:40:03.255639079 +0000 UTC m=+208.099070518" observedRunningTime="2025-12-06 10:40:04.865166078 +0000 UTC m=+209.708597517" watchObservedRunningTime="2025-12-06 10:40:04.870809949 +0000 UTC m=+209.714241388" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.115887 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.115958 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.197852 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.197918 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.260643 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.414360 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.414407 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.416655 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5gkj5"] Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.416859 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5gkj5" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="registry-server" containerID="cri-o://fcf0aa3cfc5bf9076e78306cdcc1e3b7c2e21229cbb20655bcbd3f3c8f4236ce" gracePeriod=2 Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.855878 4678 generic.go:334] "Generic (PLEG): container finished" podID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerID="169e2315fcfd980cccf4e8d3eb7be4d3024a898438c3a2a0e2bd490b1f3eba16" exitCode=0 Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.855975 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwz2" event={"ID":"512340e2-4c40-48fd-ac35-8f84f1de5c6a","Type":"ContainerDied","Data":"169e2315fcfd980cccf4e8d3eb7be4d3024a898438c3a2a0e2bd490b1f3eba16"} Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.863610 4678 generic.go:334] "Generic (PLEG): container finished" podID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerID="fcf0aa3cfc5bf9076e78306cdcc1e3b7c2e21229cbb20655bcbd3f3c8f4236ce" exitCode=0 Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.863663 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gkj5" event={"ID":"582c80c6-20e2-4a0e-9871-f1f4db232ca2","Type":"ContainerDied","Data":"fcf0aa3cfc5bf9076e78306cdcc1e3b7c2e21229cbb20655bcbd3f3c8f4236ce"} Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.873153 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfk99" event={"ID":"ebcefb39-28c6-4619-88c5-15537f549d0d","Type":"ContainerStarted","Data":"c5c87173ceaddeb726a5126f427befbc105633e33837433ce7352653e34e8ef7"} Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.910301 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rfk99" podStartSLOduration=3.668594084 podStartE2EDuration="1m1.910277365s" podCreationTimestamp="2025-12-06 10:39:04 +0000 UTC" firstStartedPulling="2025-12-06 10:39:07.119060761 +0000 UTC m=+151.962492200" lastFinishedPulling="2025-12-06 10:40:05.360744042 +0000 UTC m=+210.204175481" observedRunningTime="2025-12-06 10:40:05.899708758 +0000 UTC m=+210.743140197" watchObservedRunningTime="2025-12-06 10:40:05.910277365 +0000 UTC m=+210.753708804" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.930583 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.935602 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s8p98" podStartSLOduration=6.227627339 podStartE2EDuration="1m4.935586774s" podCreationTimestamp="2025-12-06 10:39:01 +0000 UTC" firstStartedPulling="2025-12-06 10:39:04.531126605 +0000 UTC m=+149.374558044" lastFinishedPulling="2025-12-06 10:40:03.23908604 +0000 UTC m=+208.082517479" observedRunningTime="2025-12-06 10:40:05.933182889 +0000 UTC m=+210.776614338" watchObservedRunningTime="2025-12-06 10:40:05.935586774 +0000 UTC m=+210.779018213" Dec 06 10:40:05 crc kubenswrapper[4678]: I1206 10:40:05.957221 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.056996 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-utilities\") pod \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.057084 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8hdg\" (UniqueName: \"kubernetes.io/projected/582c80c6-20e2-4a0e-9871-f1f4db232ca2-kube-api-access-l8hdg\") pod \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.057169 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-catalog-content\") pod \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\" (UID: \"582c80c6-20e2-4a0e-9871-f1f4db232ca2\") " Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.058022 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-utilities" (OuterVolumeSpecName: "utilities") pod "582c80c6-20e2-4a0e-9871-f1f4db232ca2" (UID: "582c80c6-20e2-4a0e-9871-f1f4db232ca2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.065659 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/582c80c6-20e2-4a0e-9871-f1f4db232ca2-kube-api-access-l8hdg" (OuterVolumeSpecName: "kube-api-access-l8hdg") pod "582c80c6-20e2-4a0e-9871-f1f4db232ca2" (UID: "582c80c6-20e2-4a0e-9871-f1f4db232ca2"). InnerVolumeSpecName "kube-api-access-l8hdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.121730 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "582c80c6-20e2-4a0e-9871-f1f4db232ca2" (UID: "582c80c6-20e2-4a0e-9871-f1f4db232ca2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.158370 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.158746 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/582c80c6-20e2-4a0e-9871-f1f4db232ca2-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.158761 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8hdg\" (UniqueName: \"kubernetes.io/projected/582c80c6-20e2-4a0e-9871-f1f4db232ca2-kube-api-access-l8hdg\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.163170 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-nd8l6" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="registry-server" probeResult="failure" output=< Dec 06 10:40:06 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 10:40:06 crc kubenswrapper[4678]: > Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.482634 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-czg22" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="registry-server" probeResult="failure" output=< Dec 06 10:40:06 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 10:40:06 crc kubenswrapper[4678]: > Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.898118 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwz2" event={"ID":"512340e2-4c40-48fd-ac35-8f84f1de5c6a","Type":"ContainerStarted","Data":"2d33b0fe0e384c0421598d754d84033175484ef43d9c6070302ff2d13accc4db"} Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.903959 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5gkj5" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.905600 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5gkj5" event={"ID":"582c80c6-20e2-4a0e-9871-f1f4db232ca2","Type":"ContainerDied","Data":"5241adc331e90f1036f2c7ab5fb478cb24fb4b8a1e94eeb596054c21560c4466"} Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.905658 4678 scope.go:117] "RemoveContainer" containerID="fcf0aa3cfc5bf9076e78306cdcc1e3b7c2e21229cbb20655bcbd3f3c8f4236ce" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.918736 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vkwz2" podStartSLOduration=5.523777665 podStartE2EDuration="1m5.918716207s" podCreationTimestamp="2025-12-06 10:39:01 +0000 UTC" firstStartedPulling="2025-12-06 10:39:05.860961084 +0000 UTC m=+150.704392523" lastFinishedPulling="2025-12-06 10:40:06.255899626 +0000 UTC m=+211.099331065" observedRunningTime="2025-12-06 10:40:06.91682114 +0000 UTC m=+211.760252579" watchObservedRunningTime="2025-12-06 10:40:06.918716207 +0000 UTC m=+211.762147656" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.935455 4678 scope.go:117] "RemoveContainer" containerID="fa4e40e63f9f620c299588aa86f937364405ea4b5a096e478a468a51bb3219bf" Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.944956 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5gkj5"] Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.945310 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5gkj5"] Dec 06 10:40:06 crc kubenswrapper[4678]: I1206 10:40:06.957986 4678 scope.go:117] "RemoveContainer" containerID="9142876d5fe2c4f56a1b2a7e077ec670892a4ffb41bc96e7892fd1bc20670025" Dec 06 10:40:07 crc kubenswrapper[4678]: I1206 10:40:07.482260 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" path="/var/lib/kubelet/pods/582c80c6-20e2-4a0e-9871-f1f4db232ca2/volumes" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.011767 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.013041 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.081672 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.487751 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.487859 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.548284 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.782379 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.783819 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.836078 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.987356 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.997546 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:40:12 crc kubenswrapper[4678]: I1206 10:40:12.999312 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:40:14 crc kubenswrapper[4678]: I1206 10:40:14.438198 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:40:14 crc kubenswrapper[4678]: I1206 10:40:14.438978 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:40:14 crc kubenswrapper[4678]: I1206 10:40:14.511087 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:40:15 crc kubenswrapper[4678]: I1206 10:40:15.070672 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:40:15 crc kubenswrapper[4678]: I1206 10:40:15.161372 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:40:15 crc kubenswrapper[4678]: I1206 10:40:15.216791 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9w2bp"] Dec 06 10:40:15 crc kubenswrapper[4678]: I1206 10:40:15.233395 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:40:15 crc kubenswrapper[4678]: I1206 10:40:15.459413 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:40:15 crc kubenswrapper[4678]: I1206 10:40:15.510945 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:40:16 crc kubenswrapper[4678]: I1206 10:40:16.018752 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9w2bp" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="registry-server" containerID="cri-o://a1306e68affc6040ffc13909a68ab91a75ed1ed6e6cf5895c71e20dfe3488f5d" gracePeriod=2 Dec 06 10:40:17 crc kubenswrapper[4678]: I1206 10:40:17.429430 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nd8l6"] Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.321327 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nd8l6" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="registry-server" containerID="cri-o://13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a" gracePeriod=2 Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.325144 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-czg22"] Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.325370 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-czg22" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="registry-server" containerID="cri-o://32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4" gracePeriod=2 Dec 06 10:40:20 crc kubenswrapper[4678]: E1206 10:40:20.483034 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod511df94f_24d2_43e7_a273_d0e2f2b5e6bc.slice/crio-conmon-a1306e68affc6040ffc13909a68ab91a75ed1ed6e6cf5895c71e20dfe3488f5d.scope\": RecentStats: unable to find data in memory cache]" Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.745535 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.830121 4678 generic.go:334] "Generic (PLEG): container finished" podID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerID="a1306e68affc6040ffc13909a68ab91a75ed1ed6e6cf5895c71e20dfe3488f5d" exitCode=0 Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.830155 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9w2bp" event={"ID":"511df94f-24d2-43e7-a273-d0e2f2b5e6bc","Type":"ContainerDied","Data":"a1306e68affc6040ffc13909a68ab91a75ed1ed6e6cf5895c71e20dfe3488f5d"} Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.830239 4678 scope.go:117] "RemoveContainer" containerID="a1306e68affc6040ffc13909a68ab91a75ed1ed6e6cf5895c71e20dfe3488f5d" Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.849884 4678 scope.go:117] "RemoveContainer" containerID="75545da2a222326d142ac13ac481a4bf79ab37a9810668ceb154dbca84e0b18d" Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.870039 4678 scope.go:117] "RemoveContainer" containerID="b964b66a5b4dd52274c86e257a9c768a69c0f6d509107a01ac2d92e34a868769" Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.914348 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-utilities\") pod \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.914403 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9zn7\" (UniqueName: \"kubernetes.io/projected/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-kube-api-access-t9zn7\") pod \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.914451 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-catalog-content\") pod \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\" (UID: \"511df94f-24d2-43e7-a273-d0e2f2b5e6bc\") " Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.915475 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-utilities" (OuterVolumeSpecName: "utilities") pod "511df94f-24d2-43e7-a273-d0e2f2b5e6bc" (UID: "511df94f-24d2-43e7-a273-d0e2f2b5e6bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.922259 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-kube-api-access-t9zn7" (OuterVolumeSpecName: "kube-api-access-t9zn7") pod "511df94f-24d2-43e7-a273-d0e2f2b5e6bc" (UID: "511df94f-24d2-43e7-a273-d0e2f2b5e6bc"). InnerVolumeSpecName "kube-api-access-t9zn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:40:20 crc kubenswrapper[4678]: I1206 10:40:20.975094 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "511df94f-24d2-43e7-a273-d0e2f2b5e6bc" (UID: "511df94f-24d2-43e7-a273-d0e2f2b5e6bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.015751 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.016096 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9zn7\" (UniqueName: \"kubernetes.io/projected/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-kube-api-access-t9zn7\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.016527 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/511df94f-24d2-43e7-a273-d0e2f2b5e6bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.312152 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.345265 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.422311 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm5wr\" (UniqueName: \"kubernetes.io/projected/9cd280dd-abfa-433b-bbbd-7f96542d01fc-kube-api-access-dm5wr\") pod \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.422371 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-catalog-content\") pod \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.422430 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-utilities\") pod \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\" (UID: \"9cd280dd-abfa-433b-bbbd-7f96542d01fc\") " Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.422451 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-catalog-content\") pod \"2b942ed0-624a-47e7-8801-e495760b1eda\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.423176 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-utilities" (OuterVolumeSpecName: "utilities") pod "9cd280dd-abfa-433b-bbbd-7f96542d01fc" (UID: "9cd280dd-abfa-433b-bbbd-7f96542d01fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.434481 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cd280dd-abfa-433b-bbbd-7f96542d01fc-kube-api-access-dm5wr" (OuterVolumeSpecName: "kube-api-access-dm5wr") pod "9cd280dd-abfa-433b-bbbd-7f96542d01fc" (UID: "9cd280dd-abfa-433b-bbbd-7f96542d01fc"). InnerVolumeSpecName "kube-api-access-dm5wr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.443340 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cd280dd-abfa-433b-bbbd-7f96542d01fc" (UID: "9cd280dd-abfa-433b-bbbd-7f96542d01fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.523675 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8znz9\" (UniqueName: \"kubernetes.io/projected/2b942ed0-624a-47e7-8801-e495760b1eda-kube-api-access-8znz9\") pod \"2b942ed0-624a-47e7-8801-e495760b1eda\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.523757 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-utilities\") pod \"2b942ed0-624a-47e7-8801-e495760b1eda\" (UID: \"2b942ed0-624a-47e7-8801-e495760b1eda\") " Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.524051 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm5wr\" (UniqueName: \"kubernetes.io/projected/9cd280dd-abfa-433b-bbbd-7f96542d01fc-kube-api-access-dm5wr\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.524073 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.524081 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cd280dd-abfa-433b-bbbd-7f96542d01fc-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.525394 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-utilities" (OuterVolumeSpecName: "utilities") pod "2b942ed0-624a-47e7-8801-e495760b1eda" (UID: "2b942ed0-624a-47e7-8801-e495760b1eda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.527182 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b942ed0-624a-47e7-8801-e495760b1eda-kube-api-access-8znz9" (OuterVolumeSpecName: "kube-api-access-8znz9") pod "2b942ed0-624a-47e7-8801-e495760b1eda" (UID: "2b942ed0-624a-47e7-8801-e495760b1eda"). InnerVolumeSpecName "kube-api-access-8znz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.537339 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b942ed0-624a-47e7-8801-e495760b1eda" (UID: "2b942ed0-624a-47e7-8801-e495760b1eda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.624838 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.624883 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8znz9\" (UniqueName: \"kubernetes.io/projected/2b942ed0-624a-47e7-8801-e495760b1eda-kube-api-access-8znz9\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.624895 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b942ed0-624a-47e7-8801-e495760b1eda-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.843297 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-czg22" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.843282 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czg22" event={"ID":"2b942ed0-624a-47e7-8801-e495760b1eda","Type":"ContainerDied","Data":"32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4"} Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.844676 4678 scope.go:117] "RemoveContainer" containerID="32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.843173 4678 generic.go:334] "Generic (PLEG): container finished" podID="2b942ed0-624a-47e7-8801-e495760b1eda" containerID="32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4" exitCode=0 Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.850168 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-czg22" event={"ID":"2b942ed0-624a-47e7-8801-e495760b1eda","Type":"ContainerDied","Data":"9c38062abf21476e31316b215b90a658c76367f69704c53900ae53820f756506"} Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.852105 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9w2bp" event={"ID":"511df94f-24d2-43e7-a273-d0e2f2b5e6bc","Type":"ContainerDied","Data":"7d2f14ce76cd58dca88e640266e4b39b9311edf7fb73badcb0cb0f246ac09653"} Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.852162 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9w2bp" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.855027 4678 generic.go:334] "Generic (PLEG): container finished" podID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerID="13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a" exitCode=0 Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.855075 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nd8l6" event={"ID":"9cd280dd-abfa-433b-bbbd-7f96542d01fc","Type":"ContainerDied","Data":"13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a"} Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.855094 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nd8l6" event={"ID":"9cd280dd-abfa-433b-bbbd-7f96542d01fc","Type":"ContainerDied","Data":"c21b7cac4edcb57b1f97305df2432596675d8763ef2064289e154b1dfec3e060"} Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.855167 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nd8l6" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.867743 4678 scope.go:117] "RemoveContainer" containerID="229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.879242 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nd8l6"] Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.893442 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nd8l6"] Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.909544 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9w2bp"] Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.927586 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9w2bp"] Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.936178 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-czg22"] Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.936755 4678 scope.go:117] "RemoveContainer" containerID="7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.948913 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-czg22"] Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.955730 4678 scope.go:117] "RemoveContainer" containerID="32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4" Dec 06 10:40:21 crc kubenswrapper[4678]: E1206 10:40:21.958696 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4\": container with ID starting with 32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4 not found: ID does not exist" containerID="32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.959043 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4"} err="failed to get container status \"32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4\": rpc error: code = NotFound desc = could not find container \"32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4\": container with ID starting with 32df87a3bf60ac68422e5fc217a4d269195eaebccb2d979cc8334af2f40dc1c4 not found: ID does not exist" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.959186 4678 scope.go:117] "RemoveContainer" containerID="229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600" Dec 06 10:40:21 crc kubenswrapper[4678]: E1206 10:40:21.960807 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600\": container with ID starting with 229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600 not found: ID does not exist" containerID="229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.960838 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600"} err="failed to get container status \"229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600\": rpc error: code = NotFound desc = could not find container \"229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600\": container with ID starting with 229403c59e0573cd5916b98b97527ce739f6ae5ad371adaf22d939df21e2b600 not found: ID does not exist" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.960860 4678 scope.go:117] "RemoveContainer" containerID="7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3" Dec 06 10:40:21 crc kubenswrapper[4678]: E1206 10:40:21.965187 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3\": container with ID starting with 7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3 not found: ID does not exist" containerID="7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.965231 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3"} err="failed to get container status \"7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3\": rpc error: code = NotFound desc = could not find container \"7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3\": container with ID starting with 7391fe184cebad3a42f70778af40e952d6cfba2008652adc97be4c4a7114d9b3 not found: ID does not exist" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.965262 4678 scope.go:117] "RemoveContainer" containerID="13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a" Dec 06 10:40:21 crc kubenswrapper[4678]: I1206 10:40:21.984016 4678 scope.go:117] "RemoveContainer" containerID="6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352" Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.001937 4678 scope.go:117] "RemoveContainer" containerID="c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea" Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.005250 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5vk2z"] Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.033326 4678 scope.go:117] "RemoveContainer" containerID="13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a" Dec 06 10:40:22 crc kubenswrapper[4678]: E1206 10:40:22.036656 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a\": container with ID starting with 13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a not found: ID does not exist" containerID="13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a" Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.036703 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a"} err="failed to get container status \"13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a\": rpc error: code = NotFound desc = could not find container \"13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a\": container with ID starting with 13b64390c3464be86be270eabb208ab9a961c952a5cef72eb99495331d78673a not found: ID does not exist" Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.036731 4678 scope.go:117] "RemoveContainer" containerID="6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352" Dec 06 10:40:22 crc kubenswrapper[4678]: E1206 10:40:22.037119 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352\": container with ID starting with 6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352 not found: ID does not exist" containerID="6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352" Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.037152 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352"} err="failed to get container status \"6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352\": rpc error: code = NotFound desc = could not find container \"6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352\": container with ID starting with 6a3060f67ba81246ae36adb03bca993ab950554a1c008c24159cf5dc324d4352 not found: ID does not exist" Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.037165 4678 scope.go:117] "RemoveContainer" containerID="c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea" Dec 06 10:40:22 crc kubenswrapper[4678]: E1206 10:40:22.037427 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea\": container with ID starting with c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea not found: ID does not exist" containerID="c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea" Dec 06 10:40:22 crc kubenswrapper[4678]: I1206 10:40:22.037447 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea"} err="failed to get container status \"c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea\": rpc error: code = NotFound desc = could not find container \"c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea\": container with ID starting with c58c68b67d83dd13928b3a8f42c4adf2278e4294d46f777dd241fff33aa78aea not found: ID does not exist" Dec 06 10:40:23 crc kubenswrapper[4678]: I1206 10:40:23.487645 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" path="/var/lib/kubelet/pods/2b942ed0-624a-47e7-8801-e495760b1eda/volumes" Dec 06 10:40:23 crc kubenswrapper[4678]: I1206 10:40:23.489455 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" path="/var/lib/kubelet/pods/511df94f-24d2-43e7-a273-d0e2f2b5e6bc/volumes" Dec 06 10:40:23 crc kubenswrapper[4678]: I1206 10:40:23.490360 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" path="/var/lib/kubelet/pods/9cd280dd-abfa-433b-bbbd-7f96542d01fc/volumes" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.583900 4678 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584742 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584771 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584785 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584796 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584814 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584824 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584841 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584851 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584865 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584876 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584890 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584900 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584917 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584926 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584936 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584946 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584959 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584968 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.584980 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa9995fb-c5e9-40b6-9415-14a32349cf94" containerName="pruner" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.584990 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa9995fb-c5e9-40b6-9415-14a32349cf94" containerName="pruner" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.585012 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585022 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="extract-content" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.585035 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585095 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.585118 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585129 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="extract-utilities" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585291 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b942ed0-624a-47e7-8801-e495760b1eda" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585309 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cd280dd-abfa-433b-bbbd-7f96542d01fc" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585325 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa9995fb-c5e9-40b6-9415-14a32349cf94" containerName="pruner" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585338 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="582c80c6-20e2-4a0e-9871-f1f4db232ca2" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585353 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="511df94f-24d2-43e7-a273-d0e2f2b5e6bc" containerName="registry-server" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.585868 4678 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.586060 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.586245 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5" gracePeriod=15 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.586414 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8" gracePeriod=15 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.586383 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf" gracePeriod=15 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.586541 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3" gracePeriod=15 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.586678 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a" gracePeriod=15 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.588648 4678 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.588875 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.588890 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.588907 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.588917 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.588929 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.588939 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.588950 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.588960 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.588969 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.588978 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.588991 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589000 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 10:40:26 crc kubenswrapper[4678]: E1206 10:40:26.589016 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589026 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589213 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589235 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589256 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589271 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589288 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.589299 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709383 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709448 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709512 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709547 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709577 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709608 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709637 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.709662 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810602 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810726 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810754 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810799 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810863 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810903 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810924 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810942 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810967 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.810985 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.811028 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.811054 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.811070 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.811092 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.811114 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.892942 4678 generic.go:334] "Generic (PLEG): container finished" podID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" containerID="35c971ec85e7a10afee376d8db22213cba410f3ce391bd58cc080791b688a87a" exitCode=0 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.893012 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50","Type":"ContainerDied","Data":"35c971ec85e7a10afee376d8db22213cba410f3ce391bd58cc080791b688a87a"} Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.893834 4678 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.894122 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.895694 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.896683 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.897301 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf" exitCode=0 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.897322 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a" exitCode=0 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.897330 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3" exitCode=0 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.897341 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8" exitCode=2 Dec 06 10:40:26 crc kubenswrapper[4678]: I1206 10:40:26.897380 4678 scope.go:117] "RemoveContainer" containerID="836439e37ecb7c2558c38989b2b0259804be17402e2844851ad0e94bab45325d" Dec 06 10:40:27 crc kubenswrapper[4678]: E1206 10:40:27.521987 4678 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-98427" volumeName="registry-storage" Dec 06 10:40:27 crc kubenswrapper[4678]: I1206 10:40:27.911541 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.163252 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.164396 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.332248 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kube-api-access\") pod \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.332898 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-var-lock\") pod \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.333020 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-var-lock" (OuterVolumeSpecName: "var-lock") pod "6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" (UID: "6f143aaa-75ad-47d7-a1dd-e0d5e5916d50"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.333200 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" (UID: "6f143aaa-75ad-47d7-a1dd-e0d5e5916d50"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.333376 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kubelet-dir\") pod \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\" (UID: \"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50\") " Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.334279 4678 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.334565 4678 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.343084 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" (UID: "6f143aaa-75ad-47d7-a1dd-e0d5e5916d50"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.436418 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6f143aaa-75ad-47d7-a1dd-e0d5e5916d50-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.922222 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6f143aaa-75ad-47d7-a1dd-e0d5e5916d50","Type":"ContainerDied","Data":"d3e05b1d046788ef16d4d29ae35346169c75a31634cc4e2040adf01590bdb515"} Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.923613 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3e05b1d046788ef16d4d29ae35346169c75a31634cc4e2040adf01590bdb515" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.922299 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 06 10:40:28 crc kubenswrapper[4678]: I1206 10:40:28.990921 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.057955 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.058622 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.059317 4678 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.059795 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.247680 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.247776 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.247823 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.247894 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.247920 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.248057 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.248366 4678 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.248405 4678 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.248434 4678 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.483520 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.932765 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.934387 4678 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5" exitCode=0 Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.934469 4678 scope.go:117] "RemoveContainer" containerID="afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.934586 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.935482 4678 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.935767 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.941480 4678 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.941948 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.956724 4678 scope.go:117] "RemoveContainer" containerID="4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.972729 4678 scope.go:117] "RemoveContainer" containerID="8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3" Dec 06 10:40:29 crc kubenswrapper[4678]: I1206 10:40:29.988476 4678 scope.go:117] "RemoveContainer" containerID="6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.007041 4678 scope.go:117] "RemoveContainer" containerID="8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.027600 4678 scope.go:117] "RemoveContainer" containerID="eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.059143 4678 scope.go:117] "RemoveContainer" containerID="afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.060084 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\": container with ID starting with afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf not found: ID does not exist" containerID="afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.060149 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf"} err="failed to get container status \"afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\": rpc error: code = NotFound desc = could not find container \"afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf\": container with ID starting with afaab23ee5939b0394e35a6dc79326b090a5cdd84c94e5946e03160e630d8ecf not found: ID does not exist" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.060186 4678 scope.go:117] "RemoveContainer" containerID="4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.060753 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\": container with ID starting with 4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a not found: ID does not exist" containerID="4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.060787 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a"} err="failed to get container status \"4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\": rpc error: code = NotFound desc = could not find container \"4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a\": container with ID starting with 4771a9861086b8e5b932d06dda9de9b15308d8d5e70a9f550254ce715224f50a not found: ID does not exist" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.060810 4678 scope.go:117] "RemoveContainer" containerID="8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.061104 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\": container with ID starting with 8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3 not found: ID does not exist" containerID="8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.061132 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3"} err="failed to get container status \"8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\": rpc error: code = NotFound desc = could not find container \"8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3\": container with ID starting with 8892363ff4b8a4a28bff55f0d46fadbd5a99be94dc871c6c134f4aa0f0761ee3 not found: ID does not exist" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.061147 4678 scope.go:117] "RemoveContainer" containerID="6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.061446 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\": container with ID starting with 6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8 not found: ID does not exist" containerID="6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.061597 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8"} err="failed to get container status \"6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\": rpc error: code = NotFound desc = could not find container \"6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8\": container with ID starting with 6b019c6baa279be7b23377ffdadd66252a561f244b90ee719c5601d35c3084f8 not found: ID does not exist" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.061631 4678 scope.go:117] "RemoveContainer" containerID="8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.062418 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\": container with ID starting with 8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5 not found: ID does not exist" containerID="8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.062476 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5"} err="failed to get container status \"8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\": rpc error: code = NotFound desc = could not find container \"8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5\": container with ID starting with 8cb5f36684b38ac7ef99bc3a9a9806989afa06dc0580beb9b349ee36a9c354b5 not found: ID does not exist" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.062547 4678 scope.go:117] "RemoveContainer" containerID="eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.062984 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\": container with ID starting with eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222 not found: ID does not exist" containerID="eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.063090 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222"} err="failed to get container status \"eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\": rpc error: code = NotFound desc = could not find container \"eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222\": container with ID starting with eb038157d945141ae8db26bca7cedb61579a71334b666004c0e649c1d7ad9222 not found: ID does not exist" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.834110 4678 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.834944 4678 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.835543 4678 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.835937 4678 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.836397 4678 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:30 crc kubenswrapper[4678]: I1206 10:40:30.836451 4678 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 06 10:40:30 crc kubenswrapper[4678]: E1206 10:40:30.836960 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="200ms" Dec 06 10:40:31 crc kubenswrapper[4678]: E1206 10:40:31.038066 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="400ms" Dec 06 10:40:31 crc kubenswrapper[4678]: E1206 10:40:31.440653 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="800ms" Dec 06 10:40:31 crc kubenswrapper[4678]: E1206 10:40:31.667532 4678 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:31 crc kubenswrapper[4678]: I1206 10:40:31.668409 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:31 crc kubenswrapper[4678]: E1206 10:40:31.711523 4678 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.198:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e9a2f40537ffa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 10:40:31.70916761 +0000 UTC m=+236.552599059,LastTimestamp:2025-12-06 10:40:31.70916761 +0000 UTC m=+236.552599059,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 10:40:31 crc kubenswrapper[4678]: I1206 10:40:31.956995 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a6fa2b6a180d8527eaa75423228b448b79cf1f8b9323fa69521b2125f957aa5a"} Dec 06 10:40:32 crc kubenswrapper[4678]: E1206 10:40:32.241413 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="1.6s" Dec 06 10:40:32 crc kubenswrapper[4678]: I1206 10:40:32.966342 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4"} Dec 06 10:40:32 crc kubenswrapper[4678]: I1206 10:40:32.967597 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:32 crc kubenswrapper[4678]: E1206 10:40:32.967959 4678 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:33 crc kubenswrapper[4678]: E1206 10:40:33.842641 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="3.2s" Dec 06 10:40:33 crc kubenswrapper[4678]: E1206 10:40:33.972626 4678 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:40:35 crc kubenswrapper[4678]: I1206 10:40:35.478157 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:37 crc kubenswrapper[4678]: E1206 10:40:37.044274 4678 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.198:6443: connect: connection refused" interval="6.4s" Dec 06 10:40:37 crc kubenswrapper[4678]: E1206 10:40:37.367920 4678 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.198:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e9a2f40537ffa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-06 10:40:31.70916761 +0000 UTC m=+236.552599059,LastTimestamp:2025-12-06 10:40:31.70916761 +0000 UTC m=+236.552599059,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 06 10:40:38 crc kubenswrapper[4678]: I1206 10:40:38.475839 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:38 crc kubenswrapper[4678]: I1206 10:40:38.477225 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:38 crc kubenswrapper[4678]: I1206 10:40:38.492346 4678 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:38 crc kubenswrapper[4678]: I1206 10:40:38.492372 4678 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:38 crc kubenswrapper[4678]: E1206 10:40:38.493134 4678 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:38 crc kubenswrapper[4678]: I1206 10:40:38.493580 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:39 crc kubenswrapper[4678]: I1206 10:40:39.005627 4678 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="79f47d0cae1104c3e658b950b72836e3507de59b4ba2d9c28dbaf40767a87c82" exitCode=0 Dec 06 10:40:39 crc kubenswrapper[4678]: I1206 10:40:39.005718 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"79f47d0cae1104c3e658b950b72836e3507de59b4ba2d9c28dbaf40767a87c82"} Dec 06 10:40:39 crc kubenswrapper[4678]: I1206 10:40:39.006004 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"07e908e0514616565e88fa59fcd92ce32acfd7033b23ada3dbc30d6078740b1f"} Dec 06 10:40:39 crc kubenswrapper[4678]: I1206 10:40:39.006321 4678 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:39 crc kubenswrapper[4678]: I1206 10:40:39.006334 4678 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:39 crc kubenswrapper[4678]: E1206 10:40:39.006805 4678 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:39 crc kubenswrapper[4678]: I1206 10:40:39.007093 4678 status_manager.go:851] "Failed to get status for pod" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.198:6443: connect: connection refused" Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.022453 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"73879b968b9dafc7214fddc025cffa55b024a1d9c34cb166555f7f656fdb253a"} Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.022881 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"069b866c478e0289d83c9f804d100f5899a8a89bb26af9b3857118b77101afaf"} Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.022899 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e105ebee67c0ad090ae7619103dec42a0d585d88b3a1d464be01e3a8f2e5c354"} Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.022910 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4ad38ddab70d42074283a44dc57f6735639be0432b4d9b8e832dbc74acdffd0f"} Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.029918 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.029970 4678 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630" exitCode=1 Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.030000 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630"} Dec 06 10:40:40 crc kubenswrapper[4678]: I1206 10:40:40.030657 4678 scope.go:117] "RemoveContainer" containerID="907dcc1d78d8928eee27082dbef28297696b685196ae8d83e70e7248a050e630" Dec 06 10:40:41 crc kubenswrapper[4678]: I1206 10:40:41.038153 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 06 10:40:41 crc kubenswrapper[4678]: I1206 10:40:41.038724 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ed4823c4e189ee9d19261f8e53770c85f4f53d09bd42a43886fb0c6cd12d09d8"} Dec 06 10:40:41 crc kubenswrapper[4678]: I1206 10:40:41.042397 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b67c89b7c92fb7ffb97c326a97a4531b77302114b9bd1820c36dac23ae8f35c6"} Dec 06 10:40:41 crc kubenswrapper[4678]: I1206 10:40:41.042653 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:41 crc kubenswrapper[4678]: I1206 10:40:41.042786 4678 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:41 crc kubenswrapper[4678]: I1206 10:40:41.042829 4678 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:43 crc kubenswrapper[4678]: I1206 10:40:43.494282 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:43 crc kubenswrapper[4678]: I1206 10:40:43.494566 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:43 crc kubenswrapper[4678]: I1206 10:40:43.503516 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:43 crc kubenswrapper[4678]: I1206 10:40:43.934700 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:40:43 crc kubenswrapper[4678]: I1206 10:40:43.938700 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:40:44 crc kubenswrapper[4678]: I1206 10:40:44.061143 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:40:46 crc kubenswrapper[4678]: I1206 10:40:46.055240 4678 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:46 crc kubenswrapper[4678]: I1206 10:40:46.081442 4678 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:46 crc kubenswrapper[4678]: I1206 10:40:46.081477 4678 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:46 crc kubenswrapper[4678]: I1206 10:40:46.090232 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:40:46 crc kubenswrapper[4678]: I1206 10:40:46.092595 4678 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="24031afb-10d8-41ad-8834-8d9fdec045c7" Dec 06 10:40:47 crc kubenswrapper[4678]: I1206 10:40:47.047156 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" podUID="53361e7d-fe05-4f40-a442-307cb30ecd1c" containerName="oauth-openshift" containerID="cri-o://13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d" gracePeriod=15 Dec 06 10:40:47 crc kubenswrapper[4678]: I1206 10:40:47.088122 4678 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:47 crc kubenswrapper[4678]: I1206 10:40:47.088209 4678 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="0fc07bde-92a6-4589-b434-ac369f244272" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.553305 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.754419 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-serving-cert\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755028 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-policies\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755077 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-router-certs\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755110 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-provider-selection\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755139 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-dir\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755167 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-cliconfig\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755192 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbb66\" (UniqueName: \"kubernetes.io/projected/53361e7d-fe05-4f40-a442-307cb30ecd1c-kube-api-access-pbb66\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755216 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-ocp-branding-template\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755248 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-idp-0-file-data\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755275 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-trusted-ca-bundle\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755297 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-session\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755325 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-error\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755351 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-service-ca\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.755378 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-login\") pod \"53361e7d-fe05-4f40-a442-307cb30ecd1c\" (UID: \"53361e7d-fe05-4f40-a442-307cb30ecd1c\") " Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.756929 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.758144 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.759090 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.764151 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.764310 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.764505 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.765143 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.765963 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.774289 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53361e7d-fe05-4f40-a442-307cb30ecd1c-kube-api-access-pbb66" (OuterVolumeSpecName: "kube-api-access-pbb66") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "kube-api-access-pbb66". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.774776 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.776127 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.778145 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.778408 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.779780 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "53361e7d-fe05-4f40-a442-307cb30ecd1c" (UID: "53361e7d-fe05-4f40-a442-307cb30ecd1c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856401 4678 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856444 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856458 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856470 4678 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/53361e7d-fe05-4f40-a442-307cb30ecd1c-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856480 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856505 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbb66\" (UniqueName: \"kubernetes.io/projected/53361e7d-fe05-4f40-a442-307cb30ecd1c-kube-api-access-pbb66\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856516 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856525 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856534 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856543 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856552 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856564 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856576 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:48 crc kubenswrapper[4678]: I1206 10:40:48.856588 4678 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/53361e7d-fe05-4f40-a442-307cb30ecd1c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:40:49 crc kubenswrapper[4678]: I1206 10:40:49.101163 4678 generic.go:334] "Generic (PLEG): container finished" podID="53361e7d-fe05-4f40-a442-307cb30ecd1c" containerID="13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d" exitCode=0 Dec 06 10:40:49 crc kubenswrapper[4678]: I1206 10:40:49.101221 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" event={"ID":"53361e7d-fe05-4f40-a442-307cb30ecd1c","Type":"ContainerDied","Data":"13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d"} Dec 06 10:40:49 crc kubenswrapper[4678]: I1206 10:40:49.101220 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" Dec 06 10:40:49 crc kubenswrapper[4678]: I1206 10:40:49.101280 4678 scope.go:117] "RemoveContainer" containerID="13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d" Dec 06 10:40:49 crc kubenswrapper[4678]: I1206 10:40:49.101264 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5vk2z" event={"ID":"53361e7d-fe05-4f40-a442-307cb30ecd1c","Type":"ContainerDied","Data":"f721f994eb94a9212768c6dfff627e48e4b0631edadbab6307a44d1803963e32"} Dec 06 10:40:49 crc kubenswrapper[4678]: I1206 10:40:49.126947 4678 scope.go:117] "RemoveContainer" containerID="13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d" Dec 06 10:40:49 crc kubenswrapper[4678]: E1206 10:40:49.128568 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d\": container with ID starting with 13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d not found: ID does not exist" containerID="13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d" Dec 06 10:40:49 crc kubenswrapper[4678]: I1206 10:40:49.128611 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d"} err="failed to get container status \"13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d\": rpc error: code = NotFound desc = could not find container \"13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d\": container with ID starting with 13e952246f111a1cf8684a68c827ad004ef1b18f21914bae0fa5dca59a97314d not found: ID does not exist" Dec 06 10:40:52 crc kubenswrapper[4678]: I1206 10:40:52.823028 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 06 10:40:54 crc kubenswrapper[4678]: I1206 10:40:54.817005 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 06 10:40:54 crc kubenswrapper[4678]: I1206 10:40:54.957350 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 06 10:40:55 crc kubenswrapper[4678]: I1206 10:40:55.490549 4678 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="24031afb-10d8-41ad-8834-8d9fdec045c7" Dec 06 10:40:56 crc kubenswrapper[4678]: I1206 10:40:56.066045 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 06 10:40:56 crc kubenswrapper[4678]: I1206 10:40:56.223411 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 06 10:40:57 crc kubenswrapper[4678]: I1206 10:40:57.147239 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 06 10:40:57 crc kubenswrapper[4678]: I1206 10:40:57.161692 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 06 10:40:57 crc kubenswrapper[4678]: I1206 10:40:57.167135 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 06 10:40:57 crc kubenswrapper[4678]: I1206 10:40:57.887956 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 06 10:40:58 crc kubenswrapper[4678]: I1206 10:40:58.040447 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 06 10:40:58 crc kubenswrapper[4678]: I1206 10:40:58.427994 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 06 10:40:58 crc kubenswrapper[4678]: I1206 10:40:58.505827 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 06 10:40:58 crc kubenswrapper[4678]: I1206 10:40:58.963436 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 06 10:40:59 crc kubenswrapper[4678]: I1206 10:40:59.179214 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 06 10:40:59 crc kubenswrapper[4678]: I1206 10:40:59.196907 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 06 10:40:59 crc kubenswrapper[4678]: I1206 10:40:59.219233 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 06 10:40:59 crc kubenswrapper[4678]: I1206 10:40:59.375520 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 06 10:40:59 crc kubenswrapper[4678]: I1206 10:40:59.664154 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 06 10:40:59 crc kubenswrapper[4678]: I1206 10:40:59.865133 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 06 10:40:59 crc kubenswrapper[4678]: I1206 10:40:59.927622 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.019408 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.133479 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.134818 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.289656 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.379444 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.410135 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.425948 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.487684 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.533843 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.604970 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.713061 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.817229 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.940469 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 06 10:41:00 crc kubenswrapper[4678]: I1206 10:41:00.981168 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.066598 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.068850 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.230712 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.340040 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.342147 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.364399 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.465691 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.537781 4678 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.562179 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.619097 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.624830 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.678111 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.737905 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.745260 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.753134 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.763802 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.771772 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.843013 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 06 10:41:01 crc kubenswrapper[4678]: I1206 10:41:01.933659 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.034052 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.124151 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.326360 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.352272 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.413440 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.547978 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.596263 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.610978 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.700640 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.892590 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.926951 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.948428 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 06 10:41:02 crc kubenswrapper[4678]: I1206 10:41:02.966759 4678 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.014613 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.039591 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.193014 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.263725 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.290662 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.407567 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.470289 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.489925 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.566454 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.898315 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.940966 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 06 10:41:03 crc kubenswrapper[4678]: I1206 10:41:03.956879 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.086196 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.088862 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.102922 4678 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.178529 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.189535 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.212841 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.274739 4678 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.280022 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-5vk2z"] Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.280122 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.285963 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.291758 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.307395 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.307371463 podStartE2EDuration="18.307371463s" podCreationTimestamp="2025-12-06 10:40:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:41:04.304442295 +0000 UTC m=+269.147873734" watchObservedRunningTime="2025-12-06 10:41:04.307371463 +0000 UTC m=+269.150802892" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.363829 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.367236 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.383170 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.410568 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.416290 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.434527 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.655128 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.662413 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.674730 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.711548 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.724562 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.916578 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 10:41:04 crc kubenswrapper[4678]: I1206 10:41:04.925348 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.038536 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.063845 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.198604 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.199834 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.204852 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.218530 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.228322 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.270274 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.320318 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.378982 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.409928 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.429637 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.457664 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.484524 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53361e7d-fe05-4f40-a442-307cb30ecd1c" path="/var/lib/kubelet/pods/53361e7d-fe05-4f40-a442-307cb30ecd1c/volumes" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.485694 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.546933 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.599154 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.656518 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.657640 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.661341 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.803411 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.824946 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.984715 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 10:41:05 crc kubenswrapper[4678]: I1206 10:41:05.986619 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.043984 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.058737 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.143908 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.151014 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.171516 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.191928 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.245974 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.339058 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.354683 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.434916 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.442921 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg"] Dec 06 10:41:06 crc kubenswrapper[4678]: E1206 10:41:06.443174 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" containerName="installer" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.443199 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" containerName="installer" Dec 06 10:41:06 crc kubenswrapper[4678]: E1206 10:41:06.443211 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53361e7d-fe05-4f40-a442-307cb30ecd1c" containerName="oauth-openshift" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.443220 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="53361e7d-fe05-4f40-a442-307cb30ecd1c" containerName="oauth-openshift" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.443317 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f143aaa-75ad-47d7-a1dd-e0d5e5916d50" containerName="installer" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.443336 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="53361e7d-fe05-4f40-a442-307cb30ecd1c" containerName="oauth-openshift" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.443845 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.446035 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.449752 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.450435 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.451442 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.451704 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.451928 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.452632 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.452845 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.453043 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.453248 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.453564 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.453796 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.460941 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.464332 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.467601 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.470162 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.471458 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg"] Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.529614 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.552235 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.555998 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-audit-policies\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556325 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556662 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556726 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556758 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0926d6b9-4e1c-4413-8e02-02a9d2417956-audit-dir\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556785 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7h6k\" (UniqueName: \"kubernetes.io/projected/0926d6b9-4e1c-4413-8e02-02a9d2417956-kube-api-access-h7h6k\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556813 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556842 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556875 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-login\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556900 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-error\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556925 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556958 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.556982 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-session\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.557010 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.588367 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658294 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658377 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-login\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658419 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-error\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658449 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658566 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658597 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-session\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658621 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658651 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-audit-policies\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658699 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658733 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658763 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0926d6b9-4e1c-4413-8e02-02a9d2417956-audit-dir\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658788 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7h6k\" (UniqueName: \"kubernetes.io/projected/0926d6b9-4e1c-4413-8e02-02a9d2417956-kube-api-access-h7h6k\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.658813 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.659683 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0926d6b9-4e1c-4413-8e02-02a9d2417956-audit-dir\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.660480 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-cliconfig\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.660601 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.661134 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-audit-policies\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.661279 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-service-ca\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.666136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-serving-cert\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.666731 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.668606 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-login\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.668992 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.669937 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-router-certs\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.670254 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-error\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.672303 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.676505 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.678843 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0926d6b9-4e1c-4413-8e02-02a9d2417956-v4-0-config-system-session\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.679118 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7h6k\" (UniqueName: \"kubernetes.io/projected/0926d6b9-4e1c-4413-8e02-02a9d2417956-kube-api-access-h7h6k\") pod \"oauth-openshift-54f75f9d4b-dsnqg\" (UID: \"0926d6b9-4e1c-4413-8e02-02a9d2417956\") " pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.700029 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.700038 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.714784 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.725332 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.730158 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.766860 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.875111 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.918691 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 06 10:41:06 crc kubenswrapper[4678]: I1206 10:41:06.972365 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.029247 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.057882 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.064245 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.069996 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.196141 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.299743 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.393111 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.467783 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.532315 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.641803 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.647069 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.686980 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.690942 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.693895 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.719031 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.725205 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.810305 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.848042 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.908597 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 06 10:41:07 crc kubenswrapper[4678]: I1206 10:41:07.968915 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.040867 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.143222 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.181929 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.279149 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.373056 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.500948 4678 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.501282 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4" gracePeriod=5 Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.582127 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.637742 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.738136 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.834061 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.888029 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.911001 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 06 10:41:08 crc kubenswrapper[4678]: I1206 10:41:08.939135 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.002109 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.036022 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.063481 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.196343 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.279122 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.336864 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.398646 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.406747 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.481236 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.564606 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.762613 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.795592 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.820856 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 06 10:41:09 crc kubenswrapper[4678]: E1206 10:41:09.878396 4678 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 06 10:41:09 crc kubenswrapper[4678]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-54f75f9d4b-dsnqg_openshift-authentication_0926d6b9-4e1c-4413-8e02-02a9d2417956_0(ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca): error adding pod openshift-authentication_oauth-openshift-54f75f9d4b-dsnqg to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca" Netns:"/var/run/netns/738d82e8-915d-434f-84f9-55c8e205db0e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-54f75f9d4b-dsnqg;K8S_POD_INFRA_CONTAINER_ID=ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca;K8S_POD_UID=0926d6b9-4e1c-4413-8e02-02a9d2417956" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg] networking: Multus: [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg/0926d6b9-4e1c-4413-8e02-02a9d2417956]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-54f75f9d4b-dsnqg in out of cluster comm: pod "oauth-openshift-54f75f9d4b-dsnqg" not found Dec 06 10:41:09 crc kubenswrapper[4678]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 06 10:41:09 crc kubenswrapper[4678]: > Dec 06 10:41:09 crc kubenswrapper[4678]: E1206 10:41:09.878506 4678 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 06 10:41:09 crc kubenswrapper[4678]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-54f75f9d4b-dsnqg_openshift-authentication_0926d6b9-4e1c-4413-8e02-02a9d2417956_0(ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca): error adding pod openshift-authentication_oauth-openshift-54f75f9d4b-dsnqg to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca" Netns:"/var/run/netns/738d82e8-915d-434f-84f9-55c8e205db0e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-54f75f9d4b-dsnqg;K8S_POD_INFRA_CONTAINER_ID=ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca;K8S_POD_UID=0926d6b9-4e1c-4413-8e02-02a9d2417956" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg] networking: Multus: [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg/0926d6b9-4e1c-4413-8e02-02a9d2417956]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-54f75f9d4b-dsnqg in out of cluster comm: pod "oauth-openshift-54f75f9d4b-dsnqg" not found Dec 06 10:41:09 crc kubenswrapper[4678]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 06 10:41:09 crc kubenswrapper[4678]: > pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:09 crc kubenswrapper[4678]: E1206 10:41:09.878533 4678 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 06 10:41:09 crc kubenswrapper[4678]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-54f75f9d4b-dsnqg_openshift-authentication_0926d6b9-4e1c-4413-8e02-02a9d2417956_0(ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca): error adding pod openshift-authentication_oauth-openshift-54f75f9d4b-dsnqg to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca" Netns:"/var/run/netns/738d82e8-915d-434f-84f9-55c8e205db0e" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-54f75f9d4b-dsnqg;K8S_POD_INFRA_CONTAINER_ID=ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca;K8S_POD_UID=0926d6b9-4e1c-4413-8e02-02a9d2417956" Path:"" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg] networking: Multus: [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg/0926d6b9-4e1c-4413-8e02-02a9d2417956]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-54f75f9d4b-dsnqg in out of cluster comm: pod "oauth-openshift-54f75f9d4b-dsnqg" not found Dec 06 10:41:09 crc kubenswrapper[4678]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 06 10:41:09 crc kubenswrapper[4678]: > pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:09 crc kubenswrapper[4678]: E1206 10:41:09.878596 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"oauth-openshift-54f75f9d4b-dsnqg_openshift-authentication(0926d6b9-4e1c-4413-8e02-02a9d2417956)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"oauth-openshift-54f75f9d4b-dsnqg_openshift-authentication(0926d6b9-4e1c-4413-8e02-02a9d2417956)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_oauth-openshift-54f75f9d4b-dsnqg_openshift-authentication_0926d6b9-4e1c-4413-8e02-02a9d2417956_0(ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca): error adding pod openshift-authentication_oauth-openshift-54f75f9d4b-dsnqg to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca\\\" Netns:\\\"/var/run/netns/738d82e8-915d-434f-84f9-55c8e205db0e\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication;K8S_POD_NAME=oauth-openshift-54f75f9d4b-dsnqg;K8S_POD_INFRA_CONTAINER_ID=ce6113674680ee2ca85b4d6f4b6e8bff2576d19a7934d24fa7bb015507ad32ca;K8S_POD_UID=0926d6b9-4e1c-4413-8e02-02a9d2417956\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg] networking: Multus: [openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg/0926d6b9-4e1c-4413-8e02-02a9d2417956]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod oauth-openshift-54f75f9d4b-dsnqg in out of cluster comm: pod \\\"oauth-openshift-54f75f9d4b-dsnqg\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" podUID="0926d6b9-4e1c-4413-8e02-02a9d2417956" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.922067 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.966445 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 06 10:41:09 crc kubenswrapper[4678]: I1206 10:41:09.984221 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.117527 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.262338 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.262930 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.299911 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.400907 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.423414 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.521901 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.800093 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 06 10:41:10 crc kubenswrapper[4678]: I1206 10:41:10.832265 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.182081 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.188927 4678 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.252088 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg"] Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.267734 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" event={"ID":"0926d6b9-4e1c-4413-8e02-02a9d2417956","Type":"ContainerStarted","Data":"36a8c5692609100e0f65a95cdfef71d4ac621fdc3d2e443c19a6fa13f4a9da68"} Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.284340 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.306540 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.410178 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.669864 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.737352 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.795128 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.813231 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.832435 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 06 10:41:11 crc kubenswrapper[4678]: I1206 10:41:11.852670 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.125051 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.250428 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.260414 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.277312 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" event={"ID":"0926d6b9-4e1c-4413-8e02-02a9d2417956","Type":"ContainerStarted","Data":"88a0b2cae1e7947cce46808533bcfe6a869918ae0a2807ba76ab95ff07160c65"} Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.278790 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.287801 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.291577 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.313687 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" podStartSLOduration=50.313662405 podStartE2EDuration="50.313662405s" podCreationTimestamp="2025-12-06 10:40:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:41:12.304928513 +0000 UTC m=+277.148359982" watchObservedRunningTime="2025-12-06 10:41:12.313662405 +0000 UTC m=+277.157093854" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.339194 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.675907 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.681818 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.758139 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 06 10:41:12 crc kubenswrapper[4678]: I1206 10:41:12.821528 4678 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.187179 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.210817 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.271919 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.462035 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.470138 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.609254 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.637121 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.735634 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.781617 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 10:41:13 crc kubenswrapper[4678]: I1206 10:41:13.783991 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.071010 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.071088 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.088473 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.088585 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.088739 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.088772 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.088828 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.089226 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.089229 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.089299 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.089336 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.100414 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.191315 4678 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.191391 4678 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.191405 4678 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.191421 4678 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.191432 4678 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.227992 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.289832 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.289963 4678 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4" exitCode=137 Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.290008 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.290058 4678 scope.go:117] "RemoveContainer" containerID="15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.307204 4678 scope.go:117] "RemoveContainer" containerID="15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4" Dec 06 10:41:14 crc kubenswrapper[4678]: E1206 10:41:14.307972 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4\": container with ID starting with 15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4 not found: ID does not exist" containerID="15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.308034 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4"} err="failed to get container status \"15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4\": rpc error: code = NotFound desc = could not find container \"15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4\": container with ID starting with 15dd08ae329cd5811077f8e9d84e4f8339a3a0e10ba68ad90c68f92fa2ffd9c4 not found: ID does not exist" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.337129 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.499842 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.654644 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 06 10:41:14 crc kubenswrapper[4678]: I1206 10:41:14.921354 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 06 10:41:15 crc kubenswrapper[4678]: I1206 10:41:15.197738 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 06 10:41:15 crc kubenswrapper[4678]: I1206 10:41:15.482189 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 06 10:41:15 crc kubenswrapper[4678]: I1206 10:41:15.616424 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 06 10:41:16 crc kubenswrapper[4678]: I1206 10:41:16.254922 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 06 10:41:33 crc kubenswrapper[4678]: I1206 10:41:33.407592 4678 generic.go:334] "Generic (PLEG): container finished" podID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerID="7e86c9e1dbf04cb5622d86c7e5f2fc46f0253c4be77b8f0ad683cf5210a7dd58" exitCode=0 Dec 06 10:41:33 crc kubenswrapper[4678]: I1206 10:41:33.407848 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" event={"ID":"72675578-a2c8-4f7b-a36b-99c5996703b5","Type":"ContainerDied","Data":"7e86c9e1dbf04cb5622d86c7e5f2fc46f0253c4be77b8f0ad683cf5210a7dd58"} Dec 06 10:41:33 crc kubenswrapper[4678]: I1206 10:41:33.409200 4678 scope.go:117] "RemoveContainer" containerID="7e86c9e1dbf04cb5622d86c7e5f2fc46f0253c4be77b8f0ad683cf5210a7dd58" Dec 06 10:41:34 crc kubenswrapper[4678]: I1206 10:41:34.115443 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:41:34 crc kubenswrapper[4678]: I1206 10:41:34.116007 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:41:34 crc kubenswrapper[4678]: I1206 10:41:34.416171 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" event={"ID":"72675578-a2c8-4f7b-a36b-99c5996703b5","Type":"ContainerStarted","Data":"b914e676014f514ddb00535108a028b2928e1b7a2a03a46ff1e0687d6a815a53"} Dec 06 10:41:34 crc kubenswrapper[4678]: I1206 10:41:34.416932 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:41:34 crc kubenswrapper[4678]: I1206 10:41:34.418771 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:41:35 crc kubenswrapper[4678]: I1206 10:41:35.307129 4678 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 06 10:41:38 crc kubenswrapper[4678]: I1206 10:41:38.816724 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qbk7d"] Dec 06 10:41:38 crc kubenswrapper[4678]: I1206 10:41:38.817733 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" podUID="c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" containerName="controller-manager" containerID="cri-o://b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079" gracePeriod=30 Dec 06 10:41:38 crc kubenswrapper[4678]: I1206 10:41:38.923581 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r"] Dec 06 10:41:38 crc kubenswrapper[4678]: I1206 10:41:38.923850 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" podUID="4280a0ae-d3d2-463a-9876-86967c99d560" containerName="route-controller-manager" containerID="cri-o://3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8" gracePeriod=30 Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.247751 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.313237 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.362806 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-client-ca\") pod \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.362903 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-proxy-ca-bundles\") pod \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.362955 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4m9r\" (UniqueName: \"kubernetes.io/projected/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-kube-api-access-s4m9r\") pod \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.363004 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-serving-cert\") pod \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.363072 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-config\") pod \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\" (UID: \"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.363925 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-client-ca" (OuterVolumeSpecName: "client-ca") pod "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" (UID: "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.364130 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" (UID: "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.365028 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-config" (OuterVolumeSpecName: "config") pod "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" (UID: "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.371696 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" (UID: "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.371938 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-kube-api-access-s4m9r" (OuterVolumeSpecName: "kube-api-access-s4m9r") pod "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" (UID: "c876ca41-ea42-4d6c-9f6b-ebc14e0e8732"). InnerVolumeSpecName "kube-api-access-s4m9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.445563 4678 generic.go:334] "Generic (PLEG): container finished" podID="c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" containerID="b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079" exitCode=0 Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.445650 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" event={"ID":"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732","Type":"ContainerDied","Data":"b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079"} Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.445691 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" event={"ID":"c876ca41-ea42-4d6c-9f6b-ebc14e0e8732","Type":"ContainerDied","Data":"1dd225ef85e8fda491ee1f4e768e96a5aee9a05cf28cb87c773520a78a1b82e0"} Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.445712 4678 scope.go:117] "RemoveContainer" containerID="b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.445864 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qbk7d" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.451315 4678 generic.go:334] "Generic (PLEG): container finished" podID="4280a0ae-d3d2-463a-9876-86967c99d560" containerID="3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8" exitCode=0 Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.451379 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" event={"ID":"4280a0ae-d3d2-463a-9876-86967c99d560","Type":"ContainerDied","Data":"3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8"} Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.451420 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" event={"ID":"4280a0ae-d3d2-463a-9876-86967c99d560","Type":"ContainerDied","Data":"5b27a6e629c7be4c6e311dac7e221374117bf6ceb8b9b5433b338f81bf1b09a2"} Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.451511 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.464764 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-config\") pod \"4280a0ae-d3d2-463a-9876-86967c99d560\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.464862 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-client-ca\") pod \"4280a0ae-d3d2-463a-9876-86967c99d560\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.464951 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk8x2\" (UniqueName: \"kubernetes.io/projected/4280a0ae-d3d2-463a-9876-86967c99d560-kube-api-access-pk8x2\") pod \"4280a0ae-d3d2-463a-9876-86967c99d560\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.465116 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4280a0ae-d3d2-463a-9876-86967c99d560-serving-cert\") pod \"4280a0ae-d3d2-463a-9876-86967c99d560\" (UID: \"4280a0ae-d3d2-463a-9876-86967c99d560\") " Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.465402 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.465423 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.465433 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.465446 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.465458 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4m9r\" (UniqueName: \"kubernetes.io/projected/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732-kube-api-access-s4m9r\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.466336 4678 scope.go:117] "RemoveContainer" containerID="b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.467180 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-client-ca" (OuterVolumeSpecName: "client-ca") pod "4280a0ae-d3d2-463a-9876-86967c99d560" (UID: "4280a0ae-d3d2-463a-9876-86967c99d560"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: E1206 10:41:39.467647 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079\": container with ID starting with b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079 not found: ID does not exist" containerID="b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.467688 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079"} err="failed to get container status \"b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079\": rpc error: code = NotFound desc = could not find container \"b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079\": container with ID starting with b9a1ee646045e35127471153f5819e7550ba14c8a5ece4331e1ce5bb1cece079 not found: ID does not exist" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.467720 4678 scope.go:117] "RemoveContainer" containerID="3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.468417 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-config" (OuterVolumeSpecName: "config") pod "4280a0ae-d3d2-463a-9876-86967c99d560" (UID: "4280a0ae-d3d2-463a-9876-86967c99d560"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.473379 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4280a0ae-d3d2-463a-9876-86967c99d560-kube-api-access-pk8x2" (OuterVolumeSpecName: "kube-api-access-pk8x2") pod "4280a0ae-d3d2-463a-9876-86967c99d560" (UID: "4280a0ae-d3d2-463a-9876-86967c99d560"). InnerVolumeSpecName "kube-api-access-pk8x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.474176 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4280a0ae-d3d2-463a-9876-86967c99d560-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4280a0ae-d3d2-463a-9876-86967c99d560" (UID: "4280a0ae-d3d2-463a-9876-86967c99d560"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.486291 4678 scope.go:117] "RemoveContainer" containerID="3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8" Dec 06 10:41:39 crc kubenswrapper[4678]: E1206 10:41:39.487213 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8\": container with ID starting with 3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8 not found: ID does not exist" containerID="3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.487245 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8"} err="failed to get container status \"3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8\": rpc error: code = NotFound desc = could not find container \"3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8\": container with ID starting with 3ba605c5f162a1a19f0f448122ac558768a03dadd4b903f55878e5e5a3e70dc8 not found: ID does not exist" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.499584 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qbk7d"] Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.509355 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qbk7d"] Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.539452 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-zxznx"] Dec 06 10:41:39 crc kubenswrapper[4678]: E1206 10:41:39.539957 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" containerName="controller-manager" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.540050 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" containerName="controller-manager" Dec 06 10:41:39 crc kubenswrapper[4678]: E1206 10:41:39.540120 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4280a0ae-d3d2-463a-9876-86967c99d560" containerName="route-controller-manager" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.540180 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4280a0ae-d3d2-463a-9876-86967c99d560" containerName="route-controller-manager" Dec 06 10:41:39 crc kubenswrapper[4678]: E1206 10:41:39.540245 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.540305 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.540449 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4280a0ae-d3d2-463a-9876-86967c99d560" containerName="route-controller-manager" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.540544 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" containerName="controller-manager" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.540600 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.541042 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.548051 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.548149 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.548405 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.548607 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.548939 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.548713 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.558678 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.567195 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk8x2\" (UniqueName: \"kubernetes.io/projected/4280a0ae-d3d2-463a-9876-86967c99d560-kube-api-access-pk8x2\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.567230 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4280a0ae-d3d2-463a-9876-86967c99d560-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.567262 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.567272 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4280a0ae-d3d2-463a-9876-86967c99d560-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.647822 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-zxznx"] Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.655452 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb"] Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.656301 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.668258 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgtkx\" (UniqueName: \"kubernetes.io/projected/5582963b-e0ed-4c06-b9e9-74988e881d74-kube-api-access-bgtkx\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.668312 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-proxy-ca-bundles\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.668368 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-client-ca\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.668538 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-config\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.668600 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5582963b-e0ed-4c06-b9e9-74988e881d74-serving-cert\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.701428 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb"] Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.768807 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r"] Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769536 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-client-ca\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769591 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-config\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769630 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-config\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769649 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-client-ca\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769671 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5582963b-e0ed-4c06-b9e9-74988e881d74-serving-cert\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769721 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxdpn\" (UniqueName: \"kubernetes.io/projected/3e4c816e-3208-4a12-b2a6-cafc7d2da495-kube-api-access-rxdpn\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769765 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e4c816e-3208-4a12-b2a6-cafc7d2da495-serving-cert\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769805 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgtkx\" (UniqueName: \"kubernetes.io/projected/5582963b-e0ed-4c06-b9e9-74988e881d74-kube-api-access-bgtkx\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.769827 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-proxy-ca-bundles\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.770986 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-proxy-ca-bundles\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.771294 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-client-ca\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.771327 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-config\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.773958 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5582963b-e0ed-4c06-b9e9-74988e881d74-serving-cert\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.777043 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-64p9r"] Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.786012 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgtkx\" (UniqueName: \"kubernetes.io/projected/5582963b-e0ed-4c06-b9e9-74988e881d74-kube-api-access-bgtkx\") pod \"controller-manager-cc6cd7b-zxznx\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.860262 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.871677 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-config\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.871758 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-client-ca\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.871799 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxdpn\" (UniqueName: \"kubernetes.io/projected/3e4c816e-3208-4a12-b2a6-cafc7d2da495-kube-api-access-rxdpn\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.871842 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e4c816e-3208-4a12-b2a6-cafc7d2da495-serving-cert\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.872623 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-client-ca\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.873432 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-config\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.882252 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e4c816e-3208-4a12-b2a6-cafc7d2da495-serving-cert\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.896661 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxdpn\" (UniqueName: \"kubernetes.io/projected/3e4c816e-3208-4a12-b2a6-cafc7d2da495-kube-api-access-rxdpn\") pod \"route-controller-manager-59fb6cb665-pjpjb\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:39 crc kubenswrapper[4678]: I1206 10:41:39.972677 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.117972 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-zxznx"] Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.194626 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-zxznx"] Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.223000 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb"] Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.275176 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb"] Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.461668 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" event={"ID":"5582963b-e0ed-4c06-b9e9-74988e881d74","Type":"ContainerStarted","Data":"4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5"} Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.462069 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" event={"ID":"5582963b-e0ed-4c06-b9e9-74988e881d74","Type":"ContainerStarted","Data":"87de8f33bf513e5d3d1d4913f9c27dbbc3c6ede3c0b177811cfb01f5bc8da50e"} Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.462218 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" podUID="5582963b-e0ed-4c06-b9e9-74988e881d74" containerName="controller-manager" containerID="cri-o://4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5" gracePeriod=30 Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.463037 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.467176 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" event={"ID":"3e4c816e-3208-4a12-b2a6-cafc7d2da495","Type":"ContainerStarted","Data":"76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5"} Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.467216 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" event={"ID":"3e4c816e-3208-4a12-b2a6-cafc7d2da495","Type":"ContainerStarted","Data":"0fe21a8ac3b13b54fc6801a6b9ec64bd3c7c71826c66f280bb4668601e9de153"} Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.467330 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" podUID="3e4c816e-3208-4a12-b2a6-cafc7d2da495" containerName="route-controller-manager" containerID="cri-o://76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5" gracePeriod=30 Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.467641 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.482620 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.519025 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" podStartSLOduration=1.519002436 podStartE2EDuration="1.519002436s" podCreationTimestamp="2025-12-06 10:41:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:41:40.516938183 +0000 UTC m=+305.360369622" watchObservedRunningTime="2025-12-06 10:41:40.519002436 +0000 UTC m=+305.362433875" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.519628 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" podStartSLOduration=1.519618644 podStartE2EDuration="1.519618644s" podCreationTimestamp="2025-12-06 10:41:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:41:40.503324274 +0000 UTC m=+305.346755713" watchObservedRunningTime="2025-12-06 10:41:40.519618644 +0000 UTC m=+305.363050083" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.611644 4678 patch_prober.go:28] interesting pod/route-controller-manager-59fb6cb665-pjpjb container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": read tcp 10.217.0.2:43414->10.217.0.58:8443: read: connection reset by peer" start-of-body= Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.611736 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" podUID="3e4c816e-3208-4a12-b2a6-cafc7d2da495" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": read tcp 10.217.0.2:43414->10.217.0.58:8443: read: connection reset by peer" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.830127 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.916717 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-59fb6cb665-pjpjb_3e4c816e-3208-4a12-b2a6-cafc7d2da495/route-controller-manager/0.log" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.916786 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.988450 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5582963b-e0ed-4c06-b9e9-74988e881d74-serving-cert\") pod \"5582963b-e0ed-4c06-b9e9-74988e881d74\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.988538 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-proxy-ca-bundles\") pod \"5582963b-e0ed-4c06-b9e9-74988e881d74\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.988613 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgtkx\" (UniqueName: \"kubernetes.io/projected/5582963b-e0ed-4c06-b9e9-74988e881d74-kube-api-access-bgtkx\") pod \"5582963b-e0ed-4c06-b9e9-74988e881d74\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.988642 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-config\") pod \"5582963b-e0ed-4c06-b9e9-74988e881d74\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.988662 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-client-ca\") pod \"5582963b-e0ed-4c06-b9e9-74988e881d74\" (UID: \"5582963b-e0ed-4c06-b9e9-74988e881d74\") " Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.989538 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-client-ca" (OuterVolumeSpecName: "client-ca") pod "5582963b-e0ed-4c06-b9e9-74988e881d74" (UID: "5582963b-e0ed-4c06-b9e9-74988e881d74"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.990023 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-config" (OuterVolumeSpecName: "config") pod "5582963b-e0ed-4c06-b9e9-74988e881d74" (UID: "5582963b-e0ed-4c06-b9e9-74988e881d74"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.990468 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5582963b-e0ed-4c06-b9e9-74988e881d74" (UID: "5582963b-e0ed-4c06-b9e9-74988e881d74"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.997844 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5582963b-e0ed-4c06-b9e9-74988e881d74-kube-api-access-bgtkx" (OuterVolumeSpecName: "kube-api-access-bgtkx") pod "5582963b-e0ed-4c06-b9e9-74988e881d74" (UID: "5582963b-e0ed-4c06-b9e9-74988e881d74"). InnerVolumeSpecName "kube-api-access-bgtkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:41:40 crc kubenswrapper[4678]: I1206 10:41:40.997946 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5582963b-e0ed-4c06-b9e9-74988e881d74-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5582963b-e0ed-4c06-b9e9-74988e881d74" (UID: "5582963b-e0ed-4c06-b9e9-74988e881d74"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.089745 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-config\") pod \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.089836 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-client-ca\") pod \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.089926 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxdpn\" (UniqueName: \"kubernetes.io/projected/3e4c816e-3208-4a12-b2a6-cafc7d2da495-kube-api-access-rxdpn\") pod \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.089952 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e4c816e-3208-4a12-b2a6-cafc7d2da495-serving-cert\") pod \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\" (UID: \"3e4c816e-3208-4a12-b2a6-cafc7d2da495\") " Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.090259 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgtkx\" (UniqueName: \"kubernetes.io/projected/5582963b-e0ed-4c06-b9e9-74988e881d74-kube-api-access-bgtkx\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.090281 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.090292 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.090303 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5582963b-e0ed-4c06-b9e9-74988e881d74-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.090310 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5582963b-e0ed-4c06-b9e9-74988e881d74-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.091243 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-client-ca" (OuterVolumeSpecName: "client-ca") pod "3e4c816e-3208-4a12-b2a6-cafc7d2da495" (UID: "3e4c816e-3208-4a12-b2a6-cafc7d2da495"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.091428 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-config" (OuterVolumeSpecName: "config") pod "3e4c816e-3208-4a12-b2a6-cafc7d2da495" (UID: "3e4c816e-3208-4a12-b2a6-cafc7d2da495"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.093955 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e4c816e-3208-4a12-b2a6-cafc7d2da495-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3e4c816e-3208-4a12-b2a6-cafc7d2da495" (UID: "3e4c816e-3208-4a12-b2a6-cafc7d2da495"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.095412 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e4c816e-3208-4a12-b2a6-cafc7d2da495-kube-api-access-rxdpn" (OuterVolumeSpecName: "kube-api-access-rxdpn") pod "3e4c816e-3208-4a12-b2a6-cafc7d2da495" (UID: "3e4c816e-3208-4a12-b2a6-cafc7d2da495"). InnerVolumeSpecName "kube-api-access-rxdpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.191805 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.191851 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3e4c816e-3208-4a12-b2a6-cafc7d2da495-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.191870 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxdpn\" (UniqueName: \"kubernetes.io/projected/3e4c816e-3208-4a12-b2a6-cafc7d2da495-kube-api-access-rxdpn\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.191888 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3e4c816e-3208-4a12-b2a6-cafc7d2da495-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.478180 4678 generic.go:334] "Generic (PLEG): container finished" podID="5582963b-e0ed-4c06-b9e9-74988e881d74" containerID="4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5" exitCode=0 Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.478318 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.482361 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-route-controller-manager_route-controller-manager-59fb6cb665-pjpjb_3e4c816e-3208-4a12-b2a6-cafc7d2da495/route-controller-manager/0.log" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.482412 4678 generic.go:334] "Generic (PLEG): container finished" podID="3e4c816e-3208-4a12-b2a6-cafc7d2da495" containerID="76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5" exitCode=255 Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.482479 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.491868 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4280a0ae-d3d2-463a-9876-86967c99d560" path="/var/lib/kubelet/pods/4280a0ae-d3d2-463a-9876-86967c99d560/volumes" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.492436 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c876ca41-ea42-4d6c-9f6b-ebc14e0e8732" path="/var/lib/kubelet/pods/c876ca41-ea42-4d6c-9f6b-ebc14e0e8732/volumes" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.493157 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" event={"ID":"5582963b-e0ed-4c06-b9e9-74988e881d74","Type":"ContainerDied","Data":"4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5"} Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.493190 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc6cd7b-zxznx" event={"ID":"5582963b-e0ed-4c06-b9e9-74988e881d74","Type":"ContainerDied","Data":"87de8f33bf513e5d3d1d4913f9c27dbbc3c6ede3c0b177811cfb01f5bc8da50e"} Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.493202 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" event={"ID":"3e4c816e-3208-4a12-b2a6-cafc7d2da495","Type":"ContainerDied","Data":"76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5"} Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.493215 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb" event={"ID":"3e4c816e-3208-4a12-b2a6-cafc7d2da495","Type":"ContainerDied","Data":"0fe21a8ac3b13b54fc6801a6b9ec64bd3c7c71826c66f280bb4668601e9de153"} Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.493235 4678 scope.go:117] "RemoveContainer" containerID="4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.530979 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-zxznx"] Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.535201 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-zxznx"] Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.546853 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb"] Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.549024 4678 scope.go:117] "RemoveContainer" containerID="4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5" Dec 06 10:41:41 crc kubenswrapper[4678]: E1206 10:41:41.549682 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5\": container with ID starting with 4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5 not found: ID does not exist" containerID="4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.549720 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5"} err="failed to get container status \"4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5\": rpc error: code = NotFound desc = could not find container \"4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5\": container with ID starting with 4db65bf0fc444f45b03fd9d5df85ef847ec333997d56564105aaa6b70b4079a5 not found: ID does not exist" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.549749 4678 scope.go:117] "RemoveContainer" containerID="76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.550380 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-pjpjb"] Dec 06 10:41:41 crc kubenswrapper[4678]: E1206 10:41:41.558776 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5582963b_e0ed_4c06_b9e9_74988e881d74.slice/crio-87de8f33bf513e5d3d1d4913f9c27dbbc3c6ede3c0b177811cfb01f5bc8da50e\": RecentStats: unable to find data in memory cache]" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.570939 4678 scope.go:117] "RemoveContainer" containerID="76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5" Dec 06 10:41:41 crc kubenswrapper[4678]: E1206 10:41:41.571377 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5\": container with ID starting with 76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5 not found: ID does not exist" containerID="76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.571410 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5"} err="failed to get container status \"76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5\": rpc error: code = NotFound desc = could not find container \"76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5\": container with ID starting with 76e06ef77ea8f5dbb0499c8a031bb1e42e5cb5069101902fd28ad8585bede0a5 not found: ID does not exist" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.970027 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b"] Dec 06 10:41:41 crc kubenswrapper[4678]: E1206 10:41:41.971522 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4c816e-3208-4a12-b2a6-cafc7d2da495" containerName="route-controller-manager" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.971702 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4c816e-3208-4a12-b2a6-cafc7d2da495" containerName="route-controller-manager" Dec 06 10:41:41 crc kubenswrapper[4678]: E1206 10:41:41.971865 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5582963b-e0ed-4c06-b9e9-74988e881d74" containerName="controller-manager" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.972016 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5582963b-e0ed-4c06-b9e9-74988e881d74" containerName="controller-manager" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.972306 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4c816e-3208-4a12-b2a6-cafc7d2da495" containerName="route-controller-manager" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.972442 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5582963b-e0ed-4c06-b9e9-74988e881d74" containerName="controller-manager" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.973157 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.978085 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.978695 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.979054 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.979877 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.980396 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5c8554768b-ntpbh"] Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.981629 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.987111 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b"] Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.988939 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.989380 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.989445 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.989541 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.989716 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.990093 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.990151 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.990199 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 06 10:41:41 crc kubenswrapper[4678]: I1206 10:41:41.994458 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.025857 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-client-ca\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.025922 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad172435-232c-44fe-b2e9-6cfce4d138fc-serving-cert\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.025961 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-config\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.025993 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx7wl\" (UniqueName: \"kubernetes.io/projected/ad172435-232c-44fe-b2e9-6cfce4d138fc-kube-api-access-kx7wl\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.058458 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c8554768b-ntpbh"] Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.126884 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltgd8\" (UniqueName: \"kubernetes.io/projected/751ba47e-0232-4db9-87a0-698652dee33b-kube-api-access-ltgd8\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.126928 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-proxy-ca-bundles\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.126952 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-client-ca\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.126970 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/751ba47e-0232-4db9-87a0-698652dee33b-serving-cert\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.127132 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-config\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.127226 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad172435-232c-44fe-b2e9-6cfce4d138fc-serving-cert\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.127309 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-config\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.127366 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx7wl\" (UniqueName: \"kubernetes.io/projected/ad172435-232c-44fe-b2e9-6cfce4d138fc-kube-api-access-kx7wl\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.127428 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-client-ca\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.127885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-client-ca\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.128685 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-config\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.140830 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad172435-232c-44fe-b2e9-6cfce4d138fc-serving-cert\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.147482 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx7wl\" (UniqueName: \"kubernetes.io/projected/ad172435-232c-44fe-b2e9-6cfce4d138fc-kube-api-access-kx7wl\") pod \"route-controller-manager-59f8c4dddd-dbm6b\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.228404 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltgd8\" (UniqueName: \"kubernetes.io/projected/751ba47e-0232-4db9-87a0-698652dee33b-kube-api-access-ltgd8\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.228509 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-proxy-ca-bundles\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.228549 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/751ba47e-0232-4db9-87a0-698652dee33b-serving-cert\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.228572 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-config\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.228642 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-client-ca\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.229896 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-client-ca\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.230100 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-config\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.230917 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-proxy-ca-bundles\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.234480 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/751ba47e-0232-4db9-87a0-698652dee33b-serving-cert\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.247320 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltgd8\" (UniqueName: \"kubernetes.io/projected/751ba47e-0232-4db9-87a0-698652dee33b-kube-api-access-ltgd8\") pod \"controller-manager-5c8554768b-ntpbh\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.337473 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.357595 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.792228 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b"] Dec 06 10:41:42 crc kubenswrapper[4678]: W1206 10:41:42.795623 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad172435_232c_44fe_b2e9_6cfce4d138fc.slice/crio-a23873207706b1351c1e1b913186100eca7ab0ce99da0a134bbd599cd435d60d WatchSource:0}: Error finding container a23873207706b1351c1e1b913186100eca7ab0ce99da0a134bbd599cd435d60d: Status 404 returned error can't find the container with id a23873207706b1351c1e1b913186100eca7ab0ce99da0a134bbd599cd435d60d Dec 06 10:41:42 crc kubenswrapper[4678]: I1206 10:41:42.854212 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5c8554768b-ntpbh"] Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.096587 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c8554768b-ntpbh"] Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.486430 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e4c816e-3208-4a12-b2a6-cafc7d2da495" path="/var/lib/kubelet/pods/3e4c816e-3208-4a12-b2a6-cafc7d2da495/volumes" Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.487377 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5582963b-e0ed-4c06-b9e9-74988e881d74" path="/var/lib/kubelet/pods/5582963b-e0ed-4c06-b9e9-74988e881d74/volumes" Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.500992 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" event={"ID":"ad172435-232c-44fe-b2e9-6cfce4d138fc","Type":"ContainerStarted","Data":"1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a"} Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.501064 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" event={"ID":"ad172435-232c-44fe-b2e9-6cfce4d138fc","Type":"ContainerStarted","Data":"a23873207706b1351c1e1b913186100eca7ab0ce99da0a134bbd599cd435d60d"} Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.501473 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.503317 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" event={"ID":"751ba47e-0232-4db9-87a0-698652dee33b","Type":"ContainerStarted","Data":"3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640"} Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.503362 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" event={"ID":"751ba47e-0232-4db9-87a0-698652dee33b","Type":"ContainerStarted","Data":"ed7657b4aba5768d7986e50b78162dabc49dcc6f87b9d17f41eb2f3ed78bc0d6"} Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.503864 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.509968 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.518461 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.555578 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" podStartSLOduration=3.555547865 podStartE2EDuration="3.555547865s" podCreationTimestamp="2025-12-06 10:41:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:41:43.554349878 +0000 UTC m=+308.397781337" watchObservedRunningTime="2025-12-06 10:41:43.555547865 +0000 UTC m=+308.398979304" Dec 06 10:41:43 crc kubenswrapper[4678]: I1206 10:41:43.558115 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" podStartSLOduration=3.558109052 podStartE2EDuration="3.558109052s" podCreationTimestamp="2025-12-06 10:41:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:41:43.532825321 +0000 UTC m=+308.376256770" watchObservedRunningTime="2025-12-06 10:41:43.558109052 +0000 UTC m=+308.401540491" Dec 06 10:41:44 crc kubenswrapper[4678]: I1206 10:41:44.508718 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" podUID="751ba47e-0232-4db9-87a0-698652dee33b" containerName="controller-manager" containerID="cri-o://3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640" gracePeriod=30 Dec 06 10:41:44 crc kubenswrapper[4678]: I1206 10:41:44.901453 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:44 crc kubenswrapper[4678]: I1206 10:41:44.947951 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-qjvt8"] Dec 06 10:41:44 crc kubenswrapper[4678]: E1206 10:41:44.948393 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751ba47e-0232-4db9-87a0-698652dee33b" containerName="controller-manager" Dec 06 10:41:44 crc kubenswrapper[4678]: I1206 10:41:44.948415 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="751ba47e-0232-4db9-87a0-698652dee33b" containerName="controller-manager" Dec 06 10:41:44 crc kubenswrapper[4678]: I1206 10:41:44.948599 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="751ba47e-0232-4db9-87a0-698652dee33b" containerName="controller-manager" Dec 06 10:41:44 crc kubenswrapper[4678]: I1206 10:41:44.949220 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:44 crc kubenswrapper[4678]: I1206 10:41:44.958070 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-qjvt8"] Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.068847 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/751ba47e-0232-4db9-87a0-698652dee33b-serving-cert\") pod \"751ba47e-0232-4db9-87a0-698652dee33b\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.068895 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-config\") pod \"751ba47e-0232-4db9-87a0-698652dee33b\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.068955 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-proxy-ca-bundles\") pod \"751ba47e-0232-4db9-87a0-698652dee33b\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.068972 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-client-ca\") pod \"751ba47e-0232-4db9-87a0-698652dee33b\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.068995 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltgd8\" (UniqueName: \"kubernetes.io/projected/751ba47e-0232-4db9-87a0-698652dee33b-kube-api-access-ltgd8\") pod \"751ba47e-0232-4db9-87a0-698652dee33b\" (UID: \"751ba47e-0232-4db9-87a0-698652dee33b\") " Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.069152 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-serving-cert\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.069184 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-proxy-ca-bundles\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.069218 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv5gw\" (UniqueName: \"kubernetes.io/projected/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-kube-api-access-wv5gw\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.069242 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-config\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.069265 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-client-ca\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.069847 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-client-ca" (OuterVolumeSpecName: "client-ca") pod "751ba47e-0232-4db9-87a0-698652dee33b" (UID: "751ba47e-0232-4db9-87a0-698652dee33b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.069901 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-config" (OuterVolumeSpecName: "config") pod "751ba47e-0232-4db9-87a0-698652dee33b" (UID: "751ba47e-0232-4db9-87a0-698652dee33b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.070209 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "751ba47e-0232-4db9-87a0-698652dee33b" (UID: "751ba47e-0232-4db9-87a0-698652dee33b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.077838 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751ba47e-0232-4db9-87a0-698652dee33b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "751ba47e-0232-4db9-87a0-698652dee33b" (UID: "751ba47e-0232-4db9-87a0-698652dee33b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.079698 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751ba47e-0232-4db9-87a0-698652dee33b-kube-api-access-ltgd8" (OuterVolumeSpecName: "kube-api-access-ltgd8") pod "751ba47e-0232-4db9-87a0-698652dee33b" (UID: "751ba47e-0232-4db9-87a0-698652dee33b"). InnerVolumeSpecName "kube-api-access-ltgd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.170321 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-serving-cert\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.170822 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-proxy-ca-bundles\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.171098 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv5gw\" (UniqueName: \"kubernetes.io/projected/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-kube-api-access-wv5gw\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.171294 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-config\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.171466 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-client-ca\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.171735 4678 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.171888 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.172019 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltgd8\" (UniqueName: \"kubernetes.io/projected/751ba47e-0232-4db9-87a0-698652dee33b-kube-api-access-ltgd8\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.172156 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/751ba47e-0232-4db9-87a0-698652dee33b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.172286 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/751ba47e-0232-4db9-87a0-698652dee33b-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.173051 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-proxy-ca-bundles\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.174880 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-client-ca\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.177903 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-config\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.178276 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-serving-cert\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.206529 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv5gw\" (UniqueName: \"kubernetes.io/projected/52c6f3e1-17b0-46b9-93c9-dc8736ecb671-kube-api-access-wv5gw\") pod \"controller-manager-cc6cd7b-qjvt8\" (UID: \"52c6f3e1-17b0-46b9-93c9-dc8736ecb671\") " pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.279240 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.524279 4678 generic.go:334] "Generic (PLEG): container finished" podID="751ba47e-0232-4db9-87a0-698652dee33b" containerID="3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640" exitCode=0 Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.525063 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.525424 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" event={"ID":"751ba47e-0232-4db9-87a0-698652dee33b","Type":"ContainerDied","Data":"3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640"} Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.525456 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5c8554768b-ntpbh" event={"ID":"751ba47e-0232-4db9-87a0-698652dee33b","Type":"ContainerDied","Data":"ed7657b4aba5768d7986e50b78162dabc49dcc6f87b9d17f41eb2f3ed78bc0d6"} Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.525473 4678 scope.go:117] "RemoveContainer" containerID="3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.551668 4678 scope.go:117] "RemoveContainer" containerID="3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640" Dec 06 10:41:45 crc kubenswrapper[4678]: E1206 10:41:45.554471 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640\": container with ID starting with 3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640 not found: ID does not exist" containerID="3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.554541 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640"} err="failed to get container status \"3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640\": rpc error: code = NotFound desc = could not find container \"3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640\": container with ID starting with 3e05347c5fc41175cb1e5c608a1ccc0940454b3d57deef0dc6e40de81ac7a640 not found: ID does not exist" Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.560408 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5c8554768b-ntpbh"] Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.566508 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5c8554768b-ntpbh"] Dec 06 10:41:45 crc kubenswrapper[4678]: I1206 10:41:45.576565 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cc6cd7b-qjvt8"] Dec 06 10:41:45 crc kubenswrapper[4678]: W1206 10:41:45.585694 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52c6f3e1_17b0_46b9_93c9_dc8736ecb671.slice/crio-621d2ddd283483effadbf2b0788ac72ec733e0a5215f2b410d3b483245e737de WatchSource:0}: Error finding container 621d2ddd283483effadbf2b0788ac72ec733e0a5215f2b410d3b483245e737de: Status 404 returned error can't find the container with id 621d2ddd283483effadbf2b0788ac72ec733e0a5215f2b410d3b483245e737de Dec 06 10:41:46 crc kubenswrapper[4678]: I1206 10:41:46.533955 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" event={"ID":"52c6f3e1-17b0-46b9-93c9-dc8736ecb671","Type":"ContainerStarted","Data":"50a91ed54b2f3081dafce89fa3847c5295ff7d527ce150b6cefbf542cf9850b6"} Dec 06 10:41:46 crc kubenswrapper[4678]: I1206 10:41:46.534050 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" event={"ID":"52c6f3e1-17b0-46b9-93c9-dc8736ecb671","Type":"ContainerStarted","Data":"621d2ddd283483effadbf2b0788ac72ec733e0a5215f2b410d3b483245e737de"} Dec 06 10:41:46 crc kubenswrapper[4678]: I1206 10:41:46.534286 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:46 crc kubenswrapper[4678]: I1206 10:41:46.544722 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" Dec 06 10:41:46 crc kubenswrapper[4678]: I1206 10:41:46.566877 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-cc6cd7b-qjvt8" podStartSLOduration=3.566837484 podStartE2EDuration="3.566837484s" podCreationTimestamp="2025-12-06 10:41:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:41:46.558706309 +0000 UTC m=+311.402137778" watchObservedRunningTime="2025-12-06 10:41:46.566837484 +0000 UTC m=+311.410268963" Dec 06 10:41:47 crc kubenswrapper[4678]: I1206 10:41:47.489643 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751ba47e-0232-4db9-87a0-698652dee33b" path="/var/lib/kubelet/pods/751ba47e-0232-4db9-87a0-698652dee33b/volumes" Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.950112 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkwz2"] Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.951437 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vkwz2" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="registry-server" containerID="cri-o://2d33b0fe0e384c0421598d754d84033175484ef43d9c6070302ff2d13accc4db" gracePeriod=30 Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.955093 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s8p98"] Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.955521 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s8p98" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="registry-server" containerID="cri-o://c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911" gracePeriod=30 Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.973157 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wbcwh"] Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.973542 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" containerID="cri-o://b914e676014f514ddb00535108a028b2928e1b7a2a03a46ff1e0687d6a815a53" gracePeriod=30 Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.988345 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfk99"] Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.988714 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rfk99" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="registry-server" containerID="cri-o://c5c87173ceaddeb726a5126f427befbc105633e33837433ce7352653e34e8ef7" gracePeriod=30 Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.997478 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxr6m"] Dec 06 10:42:14 crc kubenswrapper[4678]: I1206 10:42:14.998408 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pxr6m" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="registry-server" containerID="cri-o://b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381" gracePeriod=30 Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.010659 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s46l9"] Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.011884 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.031347 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s46l9"] Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.175637 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7dc2165b-f6c0-4386-b931-5e80f762dab5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.175681 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7dc2165b-f6c0-4386-b931-5e80f762dab5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.176042 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd5zt\" (UniqueName: \"kubernetes.io/projected/7dc2165b-f6c0-4386-b931-5e80f762dab5-kube-api-access-sd5zt\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: E1206 10:42:15.198875 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381 is running failed: container process not found" containerID="b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 10:42:15 crc kubenswrapper[4678]: E1206 10:42:15.199905 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381 is running failed: container process not found" containerID="b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 10:42:15 crc kubenswrapper[4678]: E1206 10:42:15.200568 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381 is running failed: container process not found" containerID="b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 10:42:15 crc kubenswrapper[4678]: E1206 10:42:15.200612 4678 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-pxr6m" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="registry-server" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.277379 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7dc2165b-f6c0-4386-b931-5e80f762dab5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.277844 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7dc2165b-f6c0-4386-b931-5e80f762dab5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.277900 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd5zt\" (UniqueName: \"kubernetes.io/projected/7dc2165b-f6c0-4386-b931-5e80f762dab5-kube-api-access-sd5zt\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.279308 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7dc2165b-f6c0-4386-b931-5e80f762dab5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.292111 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7dc2165b-f6c0-4386-b931-5e80f762dab5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.293911 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd5zt\" (UniqueName: \"kubernetes.io/projected/7dc2165b-f6c0-4386-b931-5e80f762dab5-kube-api-access-sd5zt\") pod \"marketplace-operator-79b997595-s46l9\" (UID: \"7dc2165b-f6c0-4386-b931-5e80f762dab5\") " pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.336347 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.403716 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.582896 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-utilities\") pod \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.582978 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r66k5\" (UniqueName: \"kubernetes.io/projected/29fca54e-1ec1-484b-b3c4-431ecaedd87b-kube-api-access-r66k5\") pod \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.583076 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-catalog-content\") pod \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\" (UID: \"29fca54e-1ec1-484b-b3c4-431ecaedd87b\") " Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.583945 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-utilities" (OuterVolumeSpecName: "utilities") pod "29fca54e-1ec1-484b-b3c4-431ecaedd87b" (UID: "29fca54e-1ec1-484b-b3c4-431ecaedd87b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.599946 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29fca54e-1ec1-484b-b3c4-431ecaedd87b-kube-api-access-r66k5" (OuterVolumeSpecName: "kube-api-access-r66k5") pod "29fca54e-1ec1-484b-b3c4-431ecaedd87b" (UID: "29fca54e-1ec1-484b-b3c4-431ecaedd87b"). InnerVolumeSpecName "kube-api-access-r66k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.637270 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29fca54e-1ec1-484b-b3c4-431ecaedd87b" (UID: "29fca54e-1ec1-484b-b3c4-431ecaedd87b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.685471 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.685546 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r66k5\" (UniqueName: \"kubernetes.io/projected/29fca54e-1ec1-484b-b3c4-431ecaedd87b-kube-api-access-r66k5\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.685563 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29fca54e-1ec1-484b-b3c4-431ecaedd87b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.762943 4678 generic.go:334] "Generic (PLEG): container finished" podID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerID="c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911" exitCode=0 Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.763029 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerDied","Data":"c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911"} Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.763068 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s8p98" event={"ID":"29fca54e-1ec1-484b-b3c4-431ecaedd87b","Type":"ContainerDied","Data":"53e89eb026416fae766f04161feb1875f7d4ad1bdbad3c3e74ced6b319309362"} Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.763087 4678 scope.go:117] "RemoveContainer" containerID="c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.763088 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s8p98" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.775171 4678 generic.go:334] "Generic (PLEG): container finished" podID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerID="b914e676014f514ddb00535108a028b2928e1b7a2a03a46ff1e0687d6a815a53" exitCode=0 Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.775238 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" event={"ID":"72675578-a2c8-4f7b-a36b-99c5996703b5","Type":"ContainerDied","Data":"b914e676014f514ddb00535108a028b2928e1b7a2a03a46ff1e0687d6a815a53"} Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.783029 4678 generic.go:334] "Generic (PLEG): container finished" podID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerID="2d33b0fe0e384c0421598d754d84033175484ef43d9c6070302ff2d13accc4db" exitCode=0 Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.783108 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwz2" event={"ID":"512340e2-4c40-48fd-ac35-8f84f1de5c6a","Type":"ContainerDied","Data":"2d33b0fe0e384c0421598d754d84033175484ef43d9c6070302ff2d13accc4db"} Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.812246 4678 scope.go:117] "RemoveContainer" containerID="9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.814296 4678 generic.go:334] "Generic (PLEG): container finished" podID="3e103cce-344c-4158-8661-6f695eb8fcca" containerID="b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381" exitCode=0 Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.814355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxr6m" event={"ID":"3e103cce-344c-4158-8661-6f695eb8fcca","Type":"ContainerDied","Data":"b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381"} Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.847553 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s8p98"] Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.856167 4678 generic.go:334] "Generic (PLEG): container finished" podID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerID="c5c87173ceaddeb726a5126f427befbc105633e33837433ce7352653e34e8ef7" exitCode=0 Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.856211 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfk99" event={"ID":"ebcefb39-28c6-4619-88c5-15537f549d0d","Type":"ContainerDied","Data":"c5c87173ceaddeb726a5126f427befbc105633e33837433ce7352653e34e8ef7"} Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.863030 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s8p98"] Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.882924 4678 scope.go:117] "RemoveContainer" containerID="ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.916832 4678 scope.go:117] "RemoveContainer" containerID="c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911" Dec 06 10:42:15 crc kubenswrapper[4678]: E1206 10:42:15.918640 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911\": container with ID starting with c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911 not found: ID does not exist" containerID="c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.918703 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911"} err="failed to get container status \"c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911\": rpc error: code = NotFound desc = could not find container \"c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911\": container with ID starting with c87febd92644bbfd38744ba0a8d580e7227a7f97b3178c63c1a7ac14a6d96911 not found: ID does not exist" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.918728 4678 scope.go:117] "RemoveContainer" containerID="9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0" Dec 06 10:42:15 crc kubenswrapper[4678]: E1206 10:42:15.919213 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0\": container with ID starting with 9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0 not found: ID does not exist" containerID="9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.919264 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0"} err="failed to get container status \"9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0\": rpc error: code = NotFound desc = could not find container \"9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0\": container with ID starting with 9c5d3a719da68ac56239140d48afa2c496aa4e4a04cae15ff8dd2c7773c142c0 not found: ID does not exist" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.919279 4678 scope.go:117] "RemoveContainer" containerID="ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609" Dec 06 10:42:15 crc kubenswrapper[4678]: E1206 10:42:15.919686 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609\": container with ID starting with ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609 not found: ID does not exist" containerID="ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.919737 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609"} err="failed to get container status \"ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609\": rpc error: code = NotFound desc = could not find container \"ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609\": container with ID starting with ab8ee37b2aaa9dda2efdd2482c535eec75c7c1a53ed5c76d74645e25c060b609 not found: ID does not exist" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.919755 4678 scope.go:117] "RemoveContainer" containerID="7e86c9e1dbf04cb5622d86c7e5f2fc46f0253c4be77b8f0ad683cf5210a7dd58" Dec 06 10:42:15 crc kubenswrapper[4678]: I1206 10:42:15.942231 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.025799 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.030687 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.057561 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101423 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-utilities\") pod \"ebcefb39-28c6-4619-88c5-15537f549d0d\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101488 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llgfk\" (UniqueName: \"kubernetes.io/projected/ebcefb39-28c6-4619-88c5-15537f549d0d-kube-api-access-llgfk\") pod \"ebcefb39-28c6-4619-88c5-15537f549d0d\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101522 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wng74\" (UniqueName: \"kubernetes.io/projected/72675578-a2c8-4f7b-a36b-99c5996703b5-kube-api-access-wng74\") pod \"72675578-a2c8-4f7b-a36b-99c5996703b5\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101542 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-trusted-ca\") pod \"72675578-a2c8-4f7b-a36b-99c5996703b5\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101559 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-utilities\") pod \"3e103cce-344c-4158-8661-6f695eb8fcca\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101584 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-catalog-content\") pod \"ebcefb39-28c6-4619-88c5-15537f549d0d\" (UID: \"ebcefb39-28c6-4619-88c5-15537f549d0d\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101601 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-utilities\") pod \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101637 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-operator-metrics\") pod \"72675578-a2c8-4f7b-a36b-99c5996703b5\" (UID: \"72675578-a2c8-4f7b-a36b-99c5996703b5\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101657 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-catalog-content\") pod \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101679 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zcdm\" (UniqueName: \"kubernetes.io/projected/512340e2-4c40-48fd-ac35-8f84f1de5c6a-kube-api-access-5zcdm\") pod \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\" (UID: \"512340e2-4c40-48fd-ac35-8f84f1de5c6a\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101708 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffqkr\" (UniqueName: \"kubernetes.io/projected/3e103cce-344c-4158-8661-6f695eb8fcca-kube-api-access-ffqkr\") pod \"3e103cce-344c-4158-8661-6f695eb8fcca\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.101732 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-catalog-content\") pod \"3e103cce-344c-4158-8661-6f695eb8fcca\" (UID: \"3e103cce-344c-4158-8661-6f695eb8fcca\") " Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.109899 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-utilities" (OuterVolumeSpecName: "utilities") pod "ebcefb39-28c6-4619-88c5-15537f549d0d" (UID: "ebcefb39-28c6-4619-88c5-15537f549d0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.110944 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "72675578-a2c8-4f7b-a36b-99c5996703b5" (UID: "72675578-a2c8-4f7b-a36b-99c5996703b5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.115415 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebcefb39-28c6-4619-88c5-15537f549d0d-kube-api-access-llgfk" (OuterVolumeSpecName: "kube-api-access-llgfk") pod "ebcefb39-28c6-4619-88c5-15537f549d0d" (UID: "ebcefb39-28c6-4619-88c5-15537f549d0d"). InnerVolumeSpecName "kube-api-access-llgfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.119243 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-utilities" (OuterVolumeSpecName: "utilities") pod "512340e2-4c40-48fd-ac35-8f84f1de5c6a" (UID: "512340e2-4c40-48fd-ac35-8f84f1de5c6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.120164 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-utilities" (OuterVolumeSpecName: "utilities") pod "3e103cce-344c-4158-8661-6f695eb8fcca" (UID: "3e103cce-344c-4158-8661-6f695eb8fcca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.120269 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72675578-a2c8-4f7b-a36b-99c5996703b5-kube-api-access-wng74" (OuterVolumeSpecName: "kube-api-access-wng74") pod "72675578-a2c8-4f7b-a36b-99c5996703b5" (UID: "72675578-a2c8-4f7b-a36b-99c5996703b5"). InnerVolumeSpecName "kube-api-access-wng74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.142657 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-s46l9"] Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.144025 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "72675578-a2c8-4f7b-a36b-99c5996703b5" (UID: "72675578-a2c8-4f7b-a36b-99c5996703b5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.145352 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/512340e2-4c40-48fd-ac35-8f84f1de5c6a-kube-api-access-5zcdm" (OuterVolumeSpecName: "kube-api-access-5zcdm") pod "512340e2-4c40-48fd-ac35-8f84f1de5c6a" (UID: "512340e2-4c40-48fd-ac35-8f84f1de5c6a"). InnerVolumeSpecName "kube-api-access-5zcdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.148078 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e103cce-344c-4158-8661-6f695eb8fcca-kube-api-access-ffqkr" (OuterVolumeSpecName: "kube-api-access-ffqkr") pod "3e103cce-344c-4158-8661-6f695eb8fcca" (UID: "3e103cce-344c-4158-8661-6f695eb8fcca"). InnerVolumeSpecName "kube-api-access-ffqkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.199981 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ebcefb39-28c6-4619-88c5-15537f549d0d" (UID: "ebcefb39-28c6-4619-88c5-15537f549d0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.202912 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.202948 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zcdm\" (UniqueName: \"kubernetes.io/projected/512340e2-4c40-48fd-ac35-8f84f1de5c6a-kube-api-access-5zcdm\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.202960 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffqkr\" (UniqueName: \"kubernetes.io/projected/3e103cce-344c-4158-8661-6f695eb8fcca-kube-api-access-ffqkr\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.202971 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.202987 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llgfk\" (UniqueName: \"kubernetes.io/projected/ebcefb39-28c6-4619-88c5-15537f549d0d-kube-api-access-llgfk\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.202997 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wng74\" (UniqueName: \"kubernetes.io/projected/72675578-a2c8-4f7b-a36b-99c5996703b5-kube-api-access-wng74\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.203007 4678 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/72675578-a2c8-4f7b-a36b-99c5996703b5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.203017 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.203026 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ebcefb39-28c6-4619-88c5-15537f549d0d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.203036 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.292991 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "512340e2-4c40-48fd-ac35-8f84f1de5c6a" (UID: "512340e2-4c40-48fd-ac35-8f84f1de5c6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.303744 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/512340e2-4c40-48fd-ac35-8f84f1de5c6a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.337262 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3e103cce-344c-4158-8661-6f695eb8fcca" (UID: "3e103cce-344c-4158-8661-6f695eb8fcca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.405053 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3e103cce-344c-4158-8661-6f695eb8fcca-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.863168 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" event={"ID":"72675578-a2c8-4f7b-a36b-99c5996703b5","Type":"ContainerDied","Data":"e7d2013468570424d777b54785cde7d1a85eb5578d2e3dfdfd672e2c7f3ca7bf"} Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.863563 4678 scope.go:117] "RemoveContainer" containerID="b914e676014f514ddb00535108a028b2928e1b7a2a03a46ff1e0687d6a815a53" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.863224 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-wbcwh" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.871353 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" event={"ID":"7dc2165b-f6c0-4386-b931-5e80f762dab5","Type":"ContainerStarted","Data":"73194dd15c1e3b01da875019e5d9bbacc855c00d075ac6415e29feb794290f1a"} Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.879857 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" event={"ID":"7dc2165b-f6c0-4386-b931-5e80f762dab5","Type":"ContainerStarted","Data":"236bfb89ae96997ae9d9bb9021c236e3d14a63e5e5b8a348ba68366518f8afa8"} Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.880282 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vkwz2" event={"ID":"512340e2-4c40-48fd-ac35-8f84f1de5c6a","Type":"ContainerDied","Data":"b180765307011ca9426246c6433987fb42d0cbd6e4cb63957d200453c08a6b9e"} Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.880532 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.880629 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pxr6m" event={"ID":"3e103cce-344c-4158-8661-6f695eb8fcca","Type":"ContainerDied","Data":"9ab40a73b16e556d35089072fcc79038d09cbf549afd9e7360bc9b07bf18b062"} Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.880722 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfk99" event={"ID":"ebcefb39-28c6-4619-88c5-15537f549d0d","Type":"ContainerDied","Data":"c24ec533b4bd06f94c895d60d8215f696dbc16328053d3a690ef6f7075aa57be"} Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.878108 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfk99" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.874151 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vkwz2" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.882203 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.875997 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pxr6m" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.885226 4678 scope.go:117] "RemoveContainer" containerID="2d33b0fe0e384c0421598d754d84033175484ef43d9c6070302ff2d13accc4db" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.919196 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-s46l9" podStartSLOduration=2.919171945 podStartE2EDuration="2.919171945s" podCreationTimestamp="2025-12-06 10:42:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:42:16.910997135 +0000 UTC m=+341.754428574" watchObservedRunningTime="2025-12-06 10:42:16.919171945 +0000 UTC m=+341.762603384" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.927182 4678 scope.go:117] "RemoveContainer" containerID="169e2315fcfd980cccf4e8d3eb7be4d3024a898438c3a2a0e2bd490b1f3eba16" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.944753 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wbcwh"] Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.950650 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-wbcwh"] Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.964806 4678 scope.go:117] "RemoveContainer" containerID="24ccfacc1302566a11ed8aecaed79aef646ac8adee042aa56161d25c4b25adee" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.967282 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfk99"] Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.973415 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfk99"] Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.988118 4678 scope.go:117] "RemoveContainer" containerID="b6afa3bae347cfea5f091a8436653312413775048a1151dc0d23d80ee451a381" Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.989992 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pxr6m"] Dec 06 10:42:16 crc kubenswrapper[4678]: I1206 10:42:16.995272 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pxr6m"] Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.008845 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vkwz2"] Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.012684 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vkwz2"] Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.017662 4678 scope.go:117] "RemoveContainer" containerID="998bb5ae4d5da43edc078c341ffcb4565cb2232fe2ec423fb09e3a9eae5326d9" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.036817 4678 scope.go:117] "RemoveContainer" containerID="91e7b7e6997c23038eaccb1d51b10a2b5a175d1882ee9efbb08deaa11ed8923a" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.053989 4678 scope.go:117] "RemoveContainer" containerID="c5c87173ceaddeb726a5126f427befbc105633e33837433ce7352653e34e8ef7" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.068801 4678 scope.go:117] "RemoveContainer" containerID="2e381988ada7a179645da3cb3435104419646a6ddbd9e8a7fa50611da4a2ffaf" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.091599 4678 scope.go:117] "RemoveContainer" containerID="dac94fb835d764a87d50534f5acf212c5f9ad4b4a6ab704d1724e84974387bdf" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.495853 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" path="/var/lib/kubelet/pods/29fca54e-1ec1-484b-b3c4-431ecaedd87b/volumes" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.498565 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" path="/var/lib/kubelet/pods/3e103cce-344c-4158-8661-6f695eb8fcca/volumes" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.499844 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" path="/var/lib/kubelet/pods/512340e2-4c40-48fd-ac35-8f84f1de5c6a/volumes" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.501132 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" path="/var/lib/kubelet/pods/72675578-a2c8-4f7b-a36b-99c5996703b5/volumes" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.502919 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" path="/var/lib/kubelet/pods/ebcefb39-28c6-4619-88c5-15537f549d0d/volumes" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717465 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-49glg"] Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717788 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717803 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717813 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717819 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717826 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717832 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717843 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717850 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717859 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717867 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717875 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717881 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717890 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717898 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717911 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717917 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717926 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717933 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717940 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717947 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717956 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717962 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717970 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717977 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.717984 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.717991 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="extract-utilities" Dec 06 10:42:17 crc kubenswrapper[4678]: E1206 10:42:17.718001 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.718007 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="extract-content" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.718109 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.718122 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="29fca54e-1ec1-484b-b3c4-431ecaedd87b" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.718132 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebcefb39-28c6-4619-88c5-15537f549d0d" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.718140 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e103cce-344c-4158-8661-6f695eb8fcca" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.718148 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="512340e2-4c40-48fd-ac35-8f84f1de5c6a" containerName="registry-server" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.720962 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="72675578-a2c8-4f7b-a36b-99c5996703b5" containerName="marketplace-operator" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.721591 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.724324 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.724942 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49glg"] Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.821841 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62428199-c567-4609-8f46-2f31ae74136a-catalog-content\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.821907 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62428199-c567-4609-8f46-2f31ae74136a-utilities\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.821948 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cngv\" (UniqueName: \"kubernetes.io/projected/62428199-c567-4609-8f46-2f31ae74136a-kube-api-access-9cngv\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.923719 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62428199-c567-4609-8f46-2f31ae74136a-utilities\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.923800 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cngv\" (UniqueName: \"kubernetes.io/projected/62428199-c567-4609-8f46-2f31ae74136a-kube-api-access-9cngv\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.923872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62428199-c567-4609-8f46-2f31ae74136a-catalog-content\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.924420 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62428199-c567-4609-8f46-2f31ae74136a-utilities\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.924446 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62428199-c567-4609-8f46-2f31ae74136a-catalog-content\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:17 crc kubenswrapper[4678]: I1206 10:42:17.948458 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cngv\" (UniqueName: \"kubernetes.io/projected/62428199-c567-4609-8f46-2f31ae74136a-kube-api-access-9cngv\") pod \"certified-operators-49glg\" (UID: \"62428199-c567-4609-8f46-2f31ae74136a\") " pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.036841 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.312499 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-m2vrz"] Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.313860 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.316593 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.329011 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2vrz"] Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.431706 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-utilities\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.431795 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbrt5\" (UniqueName: \"kubernetes.io/projected/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-kube-api-access-nbrt5\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.432003 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-catalog-content\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.460649 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-49glg"] Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.538647 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-catalog-content\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.539086 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-utilities\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.539267 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbrt5\" (UniqueName: \"kubernetes.io/projected/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-kube-api-access-nbrt5\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.539289 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-catalog-content\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.539712 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-utilities\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.566966 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbrt5\" (UniqueName: \"kubernetes.io/projected/2c5e1af0-281c-42a7-bddc-0c7faa9b20e5-kube-api-access-nbrt5\") pod \"community-operators-m2vrz\" (UID: \"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5\") " pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.650268 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.867724 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b"] Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.868321 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" podUID="ad172435-232c-44fe-b2e9-6cfce4d138fc" containerName="route-controller-manager" containerID="cri-o://1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a" gracePeriod=30 Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.919356 4678 generic.go:334] "Generic (PLEG): container finished" podID="62428199-c567-4609-8f46-2f31ae74136a" containerID="7b385c1b4558bbfa0f89ad2b287bd0440907f644eba80511e383a843d31bd266" exitCode=0 Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.919828 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49glg" event={"ID":"62428199-c567-4609-8f46-2f31ae74136a","Type":"ContainerDied","Data":"7b385c1b4558bbfa0f89ad2b287bd0440907f644eba80511e383a843d31bd266"} Dec 06 10:42:18 crc kubenswrapper[4678]: I1206 10:42:18.919899 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49glg" event={"ID":"62428199-c567-4609-8f46-2f31ae74136a","Type":"ContainerStarted","Data":"4bfe6aa9feb843013411a70a52fe1b2323af9c2513c63fe0d5bdc8f926bc29bf"} Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.148656 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-m2vrz"] Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.246620 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.358588 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx7wl\" (UniqueName: \"kubernetes.io/projected/ad172435-232c-44fe-b2e9-6cfce4d138fc-kube-api-access-kx7wl\") pod \"ad172435-232c-44fe-b2e9-6cfce4d138fc\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.358947 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-config\") pod \"ad172435-232c-44fe-b2e9-6cfce4d138fc\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.359063 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad172435-232c-44fe-b2e9-6cfce4d138fc-serving-cert\") pod \"ad172435-232c-44fe-b2e9-6cfce4d138fc\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.359194 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-client-ca\") pod \"ad172435-232c-44fe-b2e9-6cfce4d138fc\" (UID: \"ad172435-232c-44fe-b2e9-6cfce4d138fc\") " Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.359627 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-client-ca" (OuterVolumeSpecName: "client-ca") pod "ad172435-232c-44fe-b2e9-6cfce4d138fc" (UID: "ad172435-232c-44fe-b2e9-6cfce4d138fc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.359903 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-config" (OuterVolumeSpecName: "config") pod "ad172435-232c-44fe-b2e9-6cfce4d138fc" (UID: "ad172435-232c-44fe-b2e9-6cfce4d138fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.365140 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad172435-232c-44fe-b2e9-6cfce4d138fc-kube-api-access-kx7wl" (OuterVolumeSpecName: "kube-api-access-kx7wl") pod "ad172435-232c-44fe-b2e9-6cfce4d138fc" (UID: "ad172435-232c-44fe-b2e9-6cfce4d138fc"). InnerVolumeSpecName "kube-api-access-kx7wl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.365432 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad172435-232c-44fe-b2e9-6cfce4d138fc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ad172435-232c-44fe-b2e9-6cfce4d138fc" (UID: "ad172435-232c-44fe-b2e9-6cfce4d138fc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.460499 4678 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ad172435-232c-44fe-b2e9-6cfce4d138fc-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.460540 4678 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-client-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.460551 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx7wl\" (UniqueName: \"kubernetes.io/projected/ad172435-232c-44fe-b2e9-6cfce4d138fc-kube-api-access-kx7wl\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.460561 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad172435-232c-44fe-b2e9-6cfce4d138fc-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.927080 4678 generic.go:334] "Generic (PLEG): container finished" podID="2c5e1af0-281c-42a7-bddc-0c7faa9b20e5" containerID="a179154f8cef2bd4d9d234b9a23b17a14a83817f2852769d5e97276c37a816ed" exitCode=0 Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.927194 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2vrz" event={"ID":"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5","Type":"ContainerDied","Data":"a179154f8cef2bd4d9d234b9a23b17a14a83817f2852769d5e97276c37a816ed"} Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.927233 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2vrz" event={"ID":"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5","Type":"ContainerStarted","Data":"969a0f91aeac058b70e3260fdb4bf33afce13ab1f7d2fbf3473435173f7f2f29"} Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.930689 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49glg" event={"ID":"62428199-c567-4609-8f46-2f31ae74136a","Type":"ContainerStarted","Data":"40ead9c062ac4c2a828913c9e417b4630cd2f59fb3d56fa265e001edb6998323"} Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.933228 4678 generic.go:334] "Generic (PLEG): container finished" podID="ad172435-232c-44fe-b2e9-6cfce4d138fc" containerID="1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a" exitCode=0 Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.933281 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" event={"ID":"ad172435-232c-44fe-b2e9-6cfce4d138fc","Type":"ContainerDied","Data":"1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a"} Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.933310 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" event={"ID":"ad172435-232c-44fe-b2e9-6cfce4d138fc","Type":"ContainerDied","Data":"a23873207706b1351c1e1b913186100eca7ab0ce99da0a134bbd599cd435d60d"} Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.933330 4678 scope.go:117] "RemoveContainer" containerID="1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.933444 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.958237 4678 scope.go:117] "RemoveContainer" containerID="1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a" Dec 06 10:42:19 crc kubenswrapper[4678]: E1206 10:42:19.958940 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a\": container with ID starting with 1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a not found: ID does not exist" containerID="1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.959012 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a"} err="failed to get container status \"1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a\": rpc error: code = NotFound desc = could not find container \"1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a\": container with ID starting with 1464f0bcc08d6b54bf913063ce71e662fddaa0afa74c0bf814bc228e78e92d0a not found: ID does not exist" Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.969283 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b"] Dec 06 10:42:19 crc kubenswrapper[4678]: I1206 10:42:19.981833 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59f8c4dddd-dbm6b"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.005935 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9"] Dec 06 10:42:20 crc kubenswrapper[4678]: E1206 10:42:20.006581 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad172435-232c-44fe-b2e9-6cfce4d138fc" containerName="route-controller-manager" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.006670 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad172435-232c-44fe-b2e9-6cfce4d138fc" containerName="route-controller-manager" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.006859 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad172435-232c-44fe-b2e9-6cfce4d138fc" containerName="route-controller-manager" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.007460 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.009508 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.010211 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.010338 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.011148 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.011550 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.011943 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.027789 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.121456 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f64bv"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.137938 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f64bv"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.138071 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.143404 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.172707 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/070792fc-9d90-4dfa-a93c-7b5894aa7475-client-ca\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.172787 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbx2s\" (UniqueName: \"kubernetes.io/projected/070792fc-9d90-4dfa-a93c-7b5894aa7475-kube-api-access-lbx2s\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.172816 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/070792fc-9d90-4dfa-a93c-7b5894aa7475-config\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.172841 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/070792fc-9d90-4dfa-a93c-7b5894aa7475-serving-cert\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.274245 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbx2s\" (UniqueName: \"kubernetes.io/projected/070792fc-9d90-4dfa-a93c-7b5894aa7475-kube-api-access-lbx2s\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.274302 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/070792fc-9d90-4dfa-a93c-7b5894aa7475-config\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.274338 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fhz8\" (UniqueName: \"kubernetes.io/projected/18d2f32a-81c5-40e4-8912-6edffff17201-kube-api-access-9fhz8\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.274364 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/070792fc-9d90-4dfa-a93c-7b5894aa7475-serving-cert\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.274400 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d2f32a-81c5-40e4-8912-6edffff17201-catalog-content\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.274433 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/070792fc-9d90-4dfa-a93c-7b5894aa7475-client-ca\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.274458 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d2f32a-81c5-40e4-8912-6edffff17201-utilities\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.275527 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/070792fc-9d90-4dfa-a93c-7b5894aa7475-client-ca\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.277423 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/070792fc-9d90-4dfa-a93c-7b5894aa7475-config\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.277856 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/070792fc-9d90-4dfa-a93c-7b5894aa7475-serving-cert\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.292178 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbx2s\" (UniqueName: \"kubernetes.io/projected/070792fc-9d90-4dfa-a93c-7b5894aa7475-kube-api-access-lbx2s\") pod \"route-controller-manager-59fb6cb665-29wf9\" (UID: \"070792fc-9d90-4dfa-a93c-7b5894aa7475\") " pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.343517 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.375699 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fhz8\" (UniqueName: \"kubernetes.io/projected/18d2f32a-81c5-40e4-8912-6edffff17201-kube-api-access-9fhz8\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.375820 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d2f32a-81c5-40e4-8912-6edffff17201-catalog-content\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.375902 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d2f32a-81c5-40e4-8912-6edffff17201-utilities\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.377011 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18d2f32a-81c5-40e4-8912-6edffff17201-catalog-content\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.377182 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18d2f32a-81c5-40e4-8912-6edffff17201-utilities\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.397365 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fhz8\" (UniqueName: \"kubernetes.io/projected/18d2f32a-81c5-40e4-8912-6edffff17201-kube-api-access-9fhz8\") pod \"redhat-marketplace-f64bv\" (UID: \"18d2f32a-81c5-40e4-8912-6edffff17201\") " pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.457962 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.722509 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fc2vn"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.727791 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.731624 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.739246 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fc2vn"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.790586 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.891181 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07116017-f2bf-4443-a09b-5df280727400-catalog-content\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.891853 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07116017-f2bf-4443-a09b-5df280727400-utilities\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.891895 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbmdl\" (UniqueName: \"kubernetes.io/projected/07116017-f2bf-4443-a09b-5df280727400-kube-api-access-kbmdl\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.919321 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f64bv"] Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.941854 4678 generic.go:334] "Generic (PLEG): container finished" podID="62428199-c567-4609-8f46-2f31ae74136a" containerID="40ead9c062ac4c2a828913c9e417b4630cd2f59fb3d56fa265e001edb6998323" exitCode=0 Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.941943 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49glg" event={"ID":"62428199-c567-4609-8f46-2f31ae74136a","Type":"ContainerDied","Data":"40ead9c062ac4c2a828913c9e417b4630cd2f59fb3d56fa265e001edb6998323"} Dec 06 10:42:20 crc kubenswrapper[4678]: W1206 10:42:20.942130 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18d2f32a_81c5_40e4_8912_6edffff17201.slice/crio-d954706083f1750e19a92c1210bccc83be3a1d15e034fbddf41bdbd69af5ccf0 WatchSource:0}: Error finding container d954706083f1750e19a92c1210bccc83be3a1d15e034fbddf41bdbd69af5ccf0: Status 404 returned error can't find the container with id d954706083f1750e19a92c1210bccc83be3a1d15e034fbddf41bdbd69af5ccf0 Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.944725 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" event={"ID":"070792fc-9d90-4dfa-a93c-7b5894aa7475","Type":"ContainerStarted","Data":"0d49b74a7578a8df227bfb0ebcad65d32a4138386c7551dd3c6c7ee8de7f2630"} Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.993019 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07116017-f2bf-4443-a09b-5df280727400-catalog-content\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.993104 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07116017-f2bf-4443-a09b-5df280727400-utilities\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.993129 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbmdl\" (UniqueName: \"kubernetes.io/projected/07116017-f2bf-4443-a09b-5df280727400-kube-api-access-kbmdl\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.993880 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/07116017-f2bf-4443-a09b-5df280727400-catalog-content\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:20 crc kubenswrapper[4678]: I1206 10:42:20.994122 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/07116017-f2bf-4443-a09b-5df280727400-utilities\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.017901 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbmdl\" (UniqueName: \"kubernetes.io/projected/07116017-f2bf-4443-a09b-5df280727400-kube-api-access-kbmdl\") pod \"redhat-operators-fc2vn\" (UID: \"07116017-f2bf-4443-a09b-5df280727400\") " pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.051745 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.483877 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad172435-232c-44fe-b2e9-6cfce4d138fc" path="/var/lib/kubelet/pods/ad172435-232c-44fe-b2e9-6cfce4d138fc/volumes" Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.505702 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fc2vn"] Dec 06 10:42:21 crc kubenswrapper[4678]: W1206 10:42:21.515682 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07116017_f2bf_4443_a09b_5df280727400.slice/crio-d23db2a8fa97471141e976f08c1a0b2a76bd11a6326a4df253f74850e143025c WatchSource:0}: Error finding container d23db2a8fa97471141e976f08c1a0b2a76bd11a6326a4df253f74850e143025c: Status 404 returned error can't find the container with id d23db2a8fa97471141e976f08c1a0b2a76bd11a6326a4df253f74850e143025c Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.955040 4678 generic.go:334] "Generic (PLEG): container finished" podID="2c5e1af0-281c-42a7-bddc-0c7faa9b20e5" containerID="f2596baabfa3de4b45fb736d4dd766592a360e59d4cff6af3a2f26515db24082" exitCode=0 Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.955083 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2vrz" event={"ID":"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5","Type":"ContainerDied","Data":"f2596baabfa3de4b45fb736d4dd766592a360e59d4cff6af3a2f26515db24082"} Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.957976 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" event={"ID":"070792fc-9d90-4dfa-a93c-7b5894aa7475","Type":"ContainerStarted","Data":"78b621a99fac68737e429f6765d1eedf986d7529a50898d454e2a9aaf5b08ea8"} Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.959721 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.966059 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.966613 4678 generic.go:334] "Generic (PLEG): container finished" podID="18d2f32a-81c5-40e4-8912-6edffff17201" containerID="9108b7b061b5c4861370f4a98ea304d8d45399da554f30dbbb637ef36f9f9209" exitCode=0 Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.966668 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f64bv" event={"ID":"18d2f32a-81c5-40e4-8912-6edffff17201","Type":"ContainerDied","Data":"9108b7b061b5c4861370f4a98ea304d8d45399da554f30dbbb637ef36f9f9209"} Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.966687 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f64bv" event={"ID":"18d2f32a-81c5-40e4-8912-6edffff17201","Type":"ContainerStarted","Data":"d954706083f1750e19a92c1210bccc83be3a1d15e034fbddf41bdbd69af5ccf0"} Dec 06 10:42:21 crc kubenswrapper[4678]: I1206 10:42:21.971586 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc2vn" event={"ID":"07116017-f2bf-4443-a09b-5df280727400","Type":"ContainerStarted","Data":"d23db2a8fa97471141e976f08c1a0b2a76bd11a6326a4df253f74850e143025c"} Dec 06 10:42:22 crc kubenswrapper[4678]: I1206 10:42:22.069535 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-59fb6cb665-29wf9" podStartSLOduration=4.069507439 podStartE2EDuration="4.069507439s" podCreationTimestamp="2025-12-06 10:42:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:42:21.99425187 +0000 UTC m=+346.837683319" watchObservedRunningTime="2025-12-06 10:42:22.069507439 +0000 UTC m=+346.912938878" Dec 06 10:42:22 crc kubenswrapper[4678]: E1206 10:42:22.220129 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07116017_f2bf_4443_a09b_5df280727400.slice/crio-conmon-21bfeee881b6c2c48bce64eb5e879a9c09fa7b7cc462b924ab154e61f6c3c0d4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07116017_f2bf_4443_a09b_5df280727400.slice/crio-21bfeee881b6c2c48bce64eb5e879a9c09fa7b7cc462b924ab154e61f6c3c0d4.scope\": RecentStats: unable to find data in memory cache]" Dec 06 10:42:22 crc kubenswrapper[4678]: I1206 10:42:22.979023 4678 generic.go:334] "Generic (PLEG): container finished" podID="07116017-f2bf-4443-a09b-5df280727400" containerID="21bfeee881b6c2c48bce64eb5e879a9c09fa7b7cc462b924ab154e61f6c3c0d4" exitCode=0 Dec 06 10:42:22 crc kubenswrapper[4678]: I1206 10:42:22.979138 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc2vn" event={"ID":"07116017-f2bf-4443-a09b-5df280727400","Type":"ContainerDied","Data":"21bfeee881b6c2c48bce64eb5e879a9c09fa7b7cc462b924ab154e61f6c3c0d4"} Dec 06 10:42:22 crc kubenswrapper[4678]: I1206 10:42:22.984305 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-m2vrz" event={"ID":"2c5e1af0-281c-42a7-bddc-0c7faa9b20e5","Type":"ContainerStarted","Data":"9f4e901f82f2459a4cd106d0352674f1eb21c24a0b3aea5bd8e7a633a2fcc765"} Dec 06 10:42:22 crc kubenswrapper[4678]: I1206 10:42:22.987844 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-49glg" event={"ID":"62428199-c567-4609-8f46-2f31ae74136a","Type":"ContainerStarted","Data":"6f12c6ca1158d4ac1699f727e057038bdbe12dd6b5db2ae9222d3ea1c561465a"} Dec 06 10:42:22 crc kubenswrapper[4678]: I1206 10:42:22.994873 4678 generic.go:334] "Generic (PLEG): container finished" podID="18d2f32a-81c5-40e4-8912-6edffff17201" containerID="d03bc343a415561fdac9a17fc5788869804a27f48632e74c16a0420e9f164085" exitCode=0 Dec 06 10:42:22 crc kubenswrapper[4678]: I1206 10:42:22.995011 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f64bv" event={"ID":"18d2f32a-81c5-40e4-8912-6edffff17201","Type":"ContainerDied","Data":"d03bc343a415561fdac9a17fc5788869804a27f48632e74c16a0420e9f164085"} Dec 06 10:42:23 crc kubenswrapper[4678]: I1206 10:42:23.026658 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-49glg" podStartSLOduration=3.5913678879999997 podStartE2EDuration="6.026635237s" podCreationTimestamp="2025-12-06 10:42:17 +0000 UTC" firstStartedPulling="2025-12-06 10:42:18.922740316 +0000 UTC m=+343.766171755" lastFinishedPulling="2025-12-06 10:42:21.358007665 +0000 UTC m=+346.201439104" observedRunningTime="2025-12-06 10:42:23.022860178 +0000 UTC m=+347.866291617" watchObservedRunningTime="2025-12-06 10:42:23.026635237 +0000 UTC m=+347.870066676" Dec 06 10:42:23 crc kubenswrapper[4678]: I1206 10:42:23.060421 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-m2vrz" podStartSLOduration=2.547223372 podStartE2EDuration="5.060397445s" podCreationTimestamp="2025-12-06 10:42:18 +0000 UTC" firstStartedPulling="2025-12-06 10:42:19.929562745 +0000 UTC m=+344.772994184" lastFinishedPulling="2025-12-06 10:42:22.442736818 +0000 UTC m=+347.286168257" observedRunningTime="2025-12-06 10:42:23.056669397 +0000 UTC m=+347.900100836" watchObservedRunningTime="2025-12-06 10:42:23.060397445 +0000 UTC m=+347.903828884" Dec 06 10:42:24 crc kubenswrapper[4678]: I1206 10:42:24.005641 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f64bv" event={"ID":"18d2f32a-81c5-40e4-8912-6edffff17201","Type":"ContainerStarted","Data":"373b1979b744694d56234516d85d121be5e2c11bf5b966427fc28a9064cc068b"} Dec 06 10:42:24 crc kubenswrapper[4678]: I1206 10:42:24.009399 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc2vn" event={"ID":"07116017-f2bf-4443-a09b-5df280727400","Type":"ContainerStarted","Data":"100c2af17e600ed66f2be4fb4ec28258b0ed5ed4077771cda3012252b8ff20f6"} Dec 06 10:42:24 crc kubenswrapper[4678]: I1206 10:42:24.063733 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f64bv" podStartSLOduration=2.646869432 podStartE2EDuration="4.063700534s" podCreationTimestamp="2025-12-06 10:42:20 +0000 UTC" firstStartedPulling="2025-12-06 10:42:21.968262208 +0000 UTC m=+346.811693647" lastFinishedPulling="2025-12-06 10:42:23.38509331 +0000 UTC m=+348.228524749" observedRunningTime="2025-12-06 10:42:24.027966393 +0000 UTC m=+348.871397832" watchObservedRunningTime="2025-12-06 10:42:24.063700534 +0000 UTC m=+348.907131973" Dec 06 10:42:25 crc kubenswrapper[4678]: I1206 10:42:25.017129 4678 generic.go:334] "Generic (PLEG): container finished" podID="07116017-f2bf-4443-a09b-5df280727400" containerID="100c2af17e600ed66f2be4fb4ec28258b0ed5ed4077771cda3012252b8ff20f6" exitCode=0 Dec 06 10:42:25 crc kubenswrapper[4678]: I1206 10:42:25.017258 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc2vn" event={"ID":"07116017-f2bf-4443-a09b-5df280727400","Type":"ContainerDied","Data":"100c2af17e600ed66f2be4fb4ec28258b0ed5ed4077771cda3012252b8ff20f6"} Dec 06 10:42:27 crc kubenswrapper[4678]: I1206 10:42:27.034374 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc2vn" event={"ID":"07116017-f2bf-4443-a09b-5df280727400","Type":"ContainerStarted","Data":"bd7ae466c4ab45269bf1371bc7de9417a037306ad0c6a08c91388ef6a406a6d5"} Dec 06 10:42:27 crc kubenswrapper[4678]: I1206 10:42:27.053666 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fc2vn" podStartSLOduration=3.264646442 podStartE2EDuration="7.053644269s" podCreationTimestamp="2025-12-06 10:42:20 +0000 UTC" firstStartedPulling="2025-12-06 10:42:22.980645373 +0000 UTC m=+347.824076812" lastFinishedPulling="2025-12-06 10:42:26.7696432 +0000 UTC m=+351.613074639" observedRunningTime="2025-12-06 10:42:27.051290224 +0000 UTC m=+351.894721683" watchObservedRunningTime="2025-12-06 10:42:27.053644269 +0000 UTC m=+351.897075708" Dec 06 10:42:28 crc kubenswrapper[4678]: I1206 10:42:28.037640 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:28 crc kubenswrapper[4678]: I1206 10:42:28.038112 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:28 crc kubenswrapper[4678]: I1206 10:42:28.079799 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:28 crc kubenswrapper[4678]: I1206 10:42:28.651090 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:28 crc kubenswrapper[4678]: I1206 10:42:28.651245 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:28 crc kubenswrapper[4678]: I1206 10:42:28.714342 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:29 crc kubenswrapper[4678]: I1206 10:42:29.096920 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-m2vrz" Dec 06 10:42:29 crc kubenswrapper[4678]: I1206 10:42:29.099212 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-49glg" Dec 06 10:42:29 crc kubenswrapper[4678]: I1206 10:42:29.506063 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:42:29 crc kubenswrapper[4678]: I1206 10:42:29.506139 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:42:30 crc kubenswrapper[4678]: I1206 10:42:30.459782 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:30 crc kubenswrapper[4678]: I1206 10:42:30.459837 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:30 crc kubenswrapper[4678]: I1206 10:42:30.510656 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:31 crc kubenswrapper[4678]: I1206 10:42:31.052569 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:31 crc kubenswrapper[4678]: I1206 10:42:31.052618 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:31 crc kubenswrapper[4678]: I1206 10:42:31.115179 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f64bv" Dec 06 10:42:32 crc kubenswrapper[4678]: I1206 10:42:32.097352 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fc2vn" podUID="07116017-f2bf-4443-a09b-5df280727400" containerName="registry-server" probeResult="failure" output=< Dec 06 10:42:32 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 10:42:32 crc kubenswrapper[4678]: > Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.627259 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rmjgn"] Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.628845 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.726150 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rmjgn"] Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750405 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-registry-tls\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750503 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/283371da-7835-4867-9a7c-a58b561c563c-trusted-ca\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750527 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/283371da-7835-4867-9a7c-a58b561c563c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750556 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jdj6\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-kube-api-access-5jdj6\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750574 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/283371da-7835-4867-9a7c-a58b561c563c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750593 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-bound-sa-token\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750631 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.750649 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/283371da-7835-4867-9a7c-a58b561c563c-registry-certificates\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.788102 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.852152 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-registry-tls\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.852237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/283371da-7835-4867-9a7c-a58b561c563c-trusted-ca\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.852266 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/283371da-7835-4867-9a7c-a58b561c563c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.852294 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jdj6\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-kube-api-access-5jdj6\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.852314 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/283371da-7835-4867-9a7c-a58b561c563c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.852339 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-bound-sa-token\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.852358 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/283371da-7835-4867-9a7c-a58b561c563c-registry-certificates\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.853556 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/283371da-7835-4867-9a7c-a58b561c563c-trusted-ca\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.853931 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/283371da-7835-4867-9a7c-a58b561c563c-registry-certificates\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.854300 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/283371da-7835-4867-9a7c-a58b561c563c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.866246 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-registry-tls\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.870672 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/283371da-7835-4867-9a7c-a58b561c563c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.871616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-bound-sa-token\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.874708 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jdj6\" (UniqueName: \"kubernetes.io/projected/283371da-7835-4867-9a7c-a58b561c563c-kube-api-access-5jdj6\") pod \"image-registry-66df7c8f76-rmjgn\" (UID: \"283371da-7835-4867-9a7c-a58b561c563c\") " pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:40 crc kubenswrapper[4678]: I1206 10:42:40.947939 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:41 crc kubenswrapper[4678]: I1206 10:42:41.101813 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:41 crc kubenswrapper[4678]: I1206 10:42:41.156218 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fc2vn" Dec 06 10:42:42 crc kubenswrapper[4678]: I1206 10:42:42.166338 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-rmjgn"] Dec 06 10:42:43 crc kubenswrapper[4678]: I1206 10:42:43.134483 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" event={"ID":"283371da-7835-4867-9a7c-a58b561c563c","Type":"ContainerStarted","Data":"f13906d0867e974146e0a6510e3cc1742c79ea8fdc03ec678779608e0dce7c0e"} Dec 06 10:42:43 crc kubenswrapper[4678]: I1206 10:42:43.135017 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:42:43 crc kubenswrapper[4678]: I1206 10:42:43.135043 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" event={"ID":"283371da-7835-4867-9a7c-a58b561c563c","Type":"ContainerStarted","Data":"1dd816ee5c0b030c142cee9ffae61aa783f3ac8ae942d639f6ad69317b29d911"} Dec 06 10:42:43 crc kubenswrapper[4678]: I1206 10:42:43.174444 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" podStartSLOduration=3.174410685 podStartE2EDuration="3.174410685s" podCreationTimestamp="2025-12-06 10:42:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:42:43.166947729 +0000 UTC m=+368.010379218" watchObservedRunningTime="2025-12-06 10:42:43.174410685 +0000 UTC m=+368.017842154" Dec 06 10:42:59 crc kubenswrapper[4678]: I1206 10:42:59.506081 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:42:59 crc kubenswrapper[4678]: I1206 10:42:59.506904 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:43:00 crc kubenswrapper[4678]: I1206 10:43:00.958053 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-rmjgn" Dec 06 10:43:01 crc kubenswrapper[4678]: I1206 10:43:01.032925 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-98427"] Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.102453 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-98427" podUID="140fee33-c778-4f79-941e-430f5ae44bd9" containerName="registry" containerID="cri-o://5cb32ca746c5a658695264ebee9b62b3fbf74f9ca8b4be97e79bf1aebeb46e11" gracePeriod=30 Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.287862 4678 patch_prober.go:28] interesting pod/image-registry-697d97f7c8-98427 container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.31:5000/healthz\": dial tcp 10.217.0.31:5000: connect: connection refused" start-of-body= Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.287964 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-697d97f7c8-98427" podUID="140fee33-c778-4f79-941e-430f5ae44bd9" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.31:5000/healthz\": dial tcp 10.217.0.31:5000: connect: connection refused" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.428053 4678 generic.go:334] "Generic (PLEG): container finished" podID="140fee33-c778-4f79-941e-430f5ae44bd9" containerID="5cb32ca746c5a658695264ebee9b62b3fbf74f9ca8b4be97e79bf1aebeb46e11" exitCode=0 Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.428417 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-98427" event={"ID":"140fee33-c778-4f79-941e-430f5ae44bd9","Type":"ContainerDied","Data":"5cb32ca746c5a658695264ebee9b62b3fbf74f9ca8b4be97e79bf1aebeb46e11"} Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.627909 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.705630 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.705694 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-registry-certificates\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.705876 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.705914 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/140fee33-c778-4f79-941e-430f5ae44bd9-ca-trust-extracted\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.705960 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-trusted-ca\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.705998 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/140fee33-c778-4f79-941e-430f5ae44bd9-installation-pull-secrets\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.706018 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcgcz\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-kube-api-access-hcgcz\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.706066 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-bound-sa-token\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.706120 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-registry-tls\") pod \"140fee33-c778-4f79-941e-430f5ae44bd9\" (UID: \"140fee33-c778-4f79-941e-430f5ae44bd9\") " Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.706422 4678 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.708099 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.715755 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/140fee33-c778-4f79-941e-430f5ae44bd9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.722686 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-kube-api-access-hcgcz" (OuterVolumeSpecName: "kube-api-access-hcgcz") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "kube-api-access-hcgcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.722683 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.723356 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.723638 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.738812 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/140fee33-c778-4f79-941e-430f5ae44bd9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "140fee33-c778-4f79-941e-430f5ae44bd9" (UID: "140fee33-c778-4f79-941e-430f5ae44bd9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.807595 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/140fee33-c778-4f79-941e-430f5ae44bd9-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.807644 4678 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/140fee33-c778-4f79-941e-430f5ae44bd9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.807661 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcgcz\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-kube-api-access-hcgcz\") on node \"crc\" DevicePath \"\"" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.807673 4678 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.807685 4678 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/140fee33-c778-4f79-941e-430f5ae44bd9-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 06 10:43:26 crc kubenswrapper[4678]: I1206 10:43:26.807696 4678 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/140fee33-c778-4f79-941e-430f5ae44bd9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 06 10:43:27 crc kubenswrapper[4678]: I1206 10:43:27.438262 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-98427" event={"ID":"140fee33-c778-4f79-941e-430f5ae44bd9","Type":"ContainerDied","Data":"ea1b9c05bc813fc2bbc1e0fa541b82c55066e8918caf3d7f29b3b2e57456f97f"} Dec 06 10:43:27 crc kubenswrapper[4678]: I1206 10:43:27.438319 4678 scope.go:117] "RemoveContainer" containerID="5cb32ca746c5a658695264ebee9b62b3fbf74f9ca8b4be97e79bf1aebeb46e11" Dec 06 10:43:27 crc kubenswrapper[4678]: I1206 10:43:27.438433 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-98427" Dec 06 10:43:27 crc kubenswrapper[4678]: I1206 10:43:27.488261 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-98427"] Dec 06 10:43:27 crc kubenswrapper[4678]: I1206 10:43:27.493339 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-98427"] Dec 06 10:43:29 crc kubenswrapper[4678]: I1206 10:43:29.488260 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="140fee33-c778-4f79-941e-430f5ae44bd9" path="/var/lib/kubelet/pods/140fee33-c778-4f79-941e-430f5ae44bd9/volumes" Dec 06 10:43:29 crc kubenswrapper[4678]: I1206 10:43:29.505707 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:43:29 crc kubenswrapper[4678]: I1206 10:43:29.505796 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:43:29 crc kubenswrapper[4678]: I1206 10:43:29.505886 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:43:29 crc kubenswrapper[4678]: I1206 10:43:29.506997 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d59ce089befb53a3c54880fdb83e41fe8e063b3a198e12260fb6860590ff9b88"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 10:43:29 crc kubenswrapper[4678]: I1206 10:43:29.507100 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://d59ce089befb53a3c54880fdb83e41fe8e063b3a198e12260fb6860590ff9b88" gracePeriod=600 Dec 06 10:43:30 crc kubenswrapper[4678]: I1206 10:43:30.465875 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="d59ce089befb53a3c54880fdb83e41fe8e063b3a198e12260fb6860590ff9b88" exitCode=0 Dec 06 10:43:30 crc kubenswrapper[4678]: I1206 10:43:30.466396 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"d59ce089befb53a3c54880fdb83e41fe8e063b3a198e12260fb6860590ff9b88"} Dec 06 10:43:30 crc kubenswrapper[4678]: I1206 10:43:30.466433 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"ae3c786b6ff131ce58d580d29415be4abad2449e4cbc0a482c5cc8f0f63c97cd"} Dec 06 10:43:30 crc kubenswrapper[4678]: I1206 10:43:30.466455 4678 scope.go:117] "RemoveContainer" containerID="2f6e064d88af0ed97206e72947d657637271ae9aecfa9eb55c77c7a850158e55" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.201254 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5"] Dec 06 10:45:00 crc kubenswrapper[4678]: E1206 10:45:00.202544 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="140fee33-c778-4f79-941e-430f5ae44bd9" containerName="registry" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.202566 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="140fee33-c778-4f79-941e-430f5ae44bd9" containerName="registry" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.202723 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="140fee33-c778-4f79-941e-430f5ae44bd9" containerName="registry" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.203263 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.209723 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.210717 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.224750 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5"] Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.234675 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33764757-2f8b-4c86-982d-4e9efde21884-config-volume\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.235025 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2h25\" (UniqueName: \"kubernetes.io/projected/33764757-2f8b-4c86-982d-4e9efde21884-kube-api-access-q2h25\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.235271 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33764757-2f8b-4c86-982d-4e9efde21884-secret-volume\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.336735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33764757-2f8b-4c86-982d-4e9efde21884-config-volume\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.337133 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2h25\" (UniqueName: \"kubernetes.io/projected/33764757-2f8b-4c86-982d-4e9efde21884-kube-api-access-q2h25\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.337329 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33764757-2f8b-4c86-982d-4e9efde21884-secret-volume\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.337656 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33764757-2f8b-4c86-982d-4e9efde21884-config-volume\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.355617 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33764757-2f8b-4c86-982d-4e9efde21884-secret-volume\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.358790 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2h25\" (UniqueName: \"kubernetes.io/projected/33764757-2f8b-4c86-982d-4e9efde21884-kube-api-access-q2h25\") pod \"collect-profiles-29416965-cj9z5\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.534612 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:00 crc kubenswrapper[4678]: I1206 10:45:00.783100 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5"] Dec 06 10:45:01 crc kubenswrapper[4678]: I1206 10:45:01.215817 4678 generic.go:334] "Generic (PLEG): container finished" podID="33764757-2f8b-4c86-982d-4e9efde21884" containerID="02fcb2fdee791cd878276c1779be8ef14f72286bfb3c0db5717ea1f2e12ba311" exitCode=0 Dec 06 10:45:01 crc kubenswrapper[4678]: I1206 10:45:01.215972 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" event={"ID":"33764757-2f8b-4c86-982d-4e9efde21884","Type":"ContainerDied","Data":"02fcb2fdee791cd878276c1779be8ef14f72286bfb3c0db5717ea1f2e12ba311"} Dec 06 10:45:01 crc kubenswrapper[4678]: I1206 10:45:01.216367 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" event={"ID":"33764757-2f8b-4c86-982d-4e9efde21884","Type":"ContainerStarted","Data":"3671c900f04bce4344cfd7e83e88511802812e4f28254f5499b66469f2a2373b"} Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.445503 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.462056 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33764757-2f8b-4c86-982d-4e9efde21884-config-volume\") pod \"33764757-2f8b-4c86-982d-4e9efde21884\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.462244 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2h25\" (UniqueName: \"kubernetes.io/projected/33764757-2f8b-4c86-982d-4e9efde21884-kube-api-access-q2h25\") pod \"33764757-2f8b-4c86-982d-4e9efde21884\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.462314 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33764757-2f8b-4c86-982d-4e9efde21884-secret-volume\") pod \"33764757-2f8b-4c86-982d-4e9efde21884\" (UID: \"33764757-2f8b-4c86-982d-4e9efde21884\") " Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.464022 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/33764757-2f8b-4c86-982d-4e9efde21884-config-volume" (OuterVolumeSpecName: "config-volume") pod "33764757-2f8b-4c86-982d-4e9efde21884" (UID: "33764757-2f8b-4c86-982d-4e9efde21884"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.469829 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33764757-2f8b-4c86-982d-4e9efde21884-kube-api-access-q2h25" (OuterVolumeSpecName: "kube-api-access-q2h25") pod "33764757-2f8b-4c86-982d-4e9efde21884" (UID: "33764757-2f8b-4c86-982d-4e9efde21884"). InnerVolumeSpecName "kube-api-access-q2h25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.472065 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33764757-2f8b-4c86-982d-4e9efde21884-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "33764757-2f8b-4c86-982d-4e9efde21884" (UID: "33764757-2f8b-4c86-982d-4e9efde21884"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.565607 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/33764757-2f8b-4c86-982d-4e9efde21884-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.565689 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2h25\" (UniqueName: \"kubernetes.io/projected/33764757-2f8b-4c86-982d-4e9efde21884-kube-api-access-q2h25\") on node \"crc\" DevicePath \"\"" Dec 06 10:45:02 crc kubenswrapper[4678]: I1206 10:45:02.565716 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/33764757-2f8b-4c86-982d-4e9efde21884-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 10:45:03 crc kubenswrapper[4678]: I1206 10:45:03.240890 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" event={"ID":"33764757-2f8b-4c86-982d-4e9efde21884","Type":"ContainerDied","Data":"3671c900f04bce4344cfd7e83e88511802812e4f28254f5499b66469f2a2373b"} Dec 06 10:45:03 crc kubenswrapper[4678]: I1206 10:45:03.241031 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3671c900f04bce4344cfd7e83e88511802812e4f28254f5499b66469f2a2373b" Dec 06 10:45:03 crc kubenswrapper[4678]: I1206 10:45:03.240988 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5" Dec 06 10:45:29 crc kubenswrapper[4678]: I1206 10:45:29.507954 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:45:29 crc kubenswrapper[4678]: I1206 10:45:29.509127 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:45:59 crc kubenswrapper[4678]: I1206 10:45:59.505161 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:45:59 crc kubenswrapper[4678]: I1206 10:45:59.506225 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.505925 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.506579 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.506653 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.507569 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ae3c786b6ff131ce58d580d29415be4abad2449e4cbc0a482c5cc8f0f63c97cd"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.507694 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://ae3c786b6ff131ce58d580d29415be4abad2449e4cbc0a482c5cc8f0f63c97cd" gracePeriod=600 Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.818998 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="ae3c786b6ff131ce58d580d29415be4abad2449e4cbc0a482c5cc8f0f63c97cd" exitCode=0 Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.819068 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"ae3c786b6ff131ce58d580d29415be4abad2449e4cbc0a482c5cc8f0f63c97cd"} Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.819549 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"74cc7443fbc9fd2e3762e8b7e18efbd09edd85649bb12e6ad42344d0b88b6e39"} Dec 06 10:46:29 crc kubenswrapper[4678]: I1206 10:46:29.819580 4678 scope.go:117] "RemoveContainer" containerID="d59ce089befb53a3c54880fdb83e41fe8e063b3a198e12260fb6860590ff9b88" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.902159 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8889z"] Dec 06 10:48:00 crc kubenswrapper[4678]: E1206 10:48:00.903102 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33764757-2f8b-4c86-982d-4e9efde21884" containerName="collect-profiles" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.903114 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="33764757-2f8b-4c86-982d-4e9efde21884" containerName="collect-profiles" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.903209 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="33764757-2f8b-4c86-982d-4e9efde21884" containerName="collect-profiles" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.903642 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.905803 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.906051 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.910660 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8889z"] Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.911153 4678 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-492l4" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.916352 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ckqhf"] Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.917327 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-ckqhf" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.923478 4678 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-6sgrg" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.939192 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ckqhf"] Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.945349 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rhwnp"] Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.946198 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.968597 4678 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-zgn2l" Dec 06 10:48:00 crc kubenswrapper[4678]: I1206 10:48:00.970687 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rhwnp"] Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.083456 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsqvg\" (UniqueName: \"kubernetes.io/projected/048fff94-2a2e-4819-af68-f3e8b2f2fa3f-kube-api-access-wsqvg\") pod \"cert-manager-5b446d88c5-ckqhf\" (UID: \"048fff94-2a2e-4819-af68-f3e8b2f2fa3f\") " pod="cert-manager/cert-manager-5b446d88c5-ckqhf" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.083541 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hvnd\" (UniqueName: \"kubernetes.io/projected/5e3d90b8-1cbb-4329-8d55-911e6dad3243-kube-api-access-7hvnd\") pod \"cert-manager-webhook-5655c58dd6-rhwnp\" (UID: \"5e3d90b8-1cbb-4329-8d55-911e6dad3243\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.083586 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686t8\" (UniqueName: \"kubernetes.io/projected/882b9c0d-2380-42b7-b13f-480a4f53a258-kube-api-access-686t8\") pod \"cert-manager-cainjector-7f985d654d-8889z\" (UID: \"882b9c0d-2380-42b7-b13f-480a4f53a258\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.184959 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsqvg\" (UniqueName: \"kubernetes.io/projected/048fff94-2a2e-4819-af68-f3e8b2f2fa3f-kube-api-access-wsqvg\") pod \"cert-manager-5b446d88c5-ckqhf\" (UID: \"048fff94-2a2e-4819-af68-f3e8b2f2fa3f\") " pod="cert-manager/cert-manager-5b446d88c5-ckqhf" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.185013 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hvnd\" (UniqueName: \"kubernetes.io/projected/5e3d90b8-1cbb-4329-8d55-911e6dad3243-kube-api-access-7hvnd\") pod \"cert-manager-webhook-5655c58dd6-rhwnp\" (UID: \"5e3d90b8-1cbb-4329-8d55-911e6dad3243\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.185041 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686t8\" (UniqueName: \"kubernetes.io/projected/882b9c0d-2380-42b7-b13f-480a4f53a258-kube-api-access-686t8\") pod \"cert-manager-cainjector-7f985d654d-8889z\" (UID: \"882b9c0d-2380-42b7-b13f-480a4f53a258\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.206304 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsqvg\" (UniqueName: \"kubernetes.io/projected/048fff94-2a2e-4819-af68-f3e8b2f2fa3f-kube-api-access-wsqvg\") pod \"cert-manager-5b446d88c5-ckqhf\" (UID: \"048fff94-2a2e-4819-af68-f3e8b2f2fa3f\") " pod="cert-manager/cert-manager-5b446d88c5-ckqhf" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.207231 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hvnd\" (UniqueName: \"kubernetes.io/projected/5e3d90b8-1cbb-4329-8d55-911e6dad3243-kube-api-access-7hvnd\") pod \"cert-manager-webhook-5655c58dd6-rhwnp\" (UID: \"5e3d90b8-1cbb-4329-8d55-911e6dad3243\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.210807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686t8\" (UniqueName: \"kubernetes.io/projected/882b9c0d-2380-42b7-b13f-480a4f53a258-kube-api-access-686t8\") pod \"cert-manager-cainjector-7f985d654d-8889z\" (UID: \"882b9c0d-2380-42b7-b13f-480a4f53a258\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.220607 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.236931 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-ckqhf" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.258586 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.506442 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-8889z"] Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.519618 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.559010 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ckqhf"] Dec 06 10:48:01 crc kubenswrapper[4678]: W1206 10:48:01.566715 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod048fff94_2a2e_4819_af68_f3e8b2f2fa3f.slice/crio-ad328b54ff67c55cd0d3600a0b3e878bcf0e84f743f7363ad5dbf876f2e8e8c9 WatchSource:0}: Error finding container ad328b54ff67c55cd0d3600a0b3e878bcf0e84f743f7363ad5dbf876f2e8e8c9: Status 404 returned error can't find the container with id ad328b54ff67c55cd0d3600a0b3e878bcf0e84f743f7363ad5dbf876f2e8e8c9 Dec 06 10:48:01 crc kubenswrapper[4678]: I1206 10:48:01.625529 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-rhwnp"] Dec 06 10:48:02 crc kubenswrapper[4678]: I1206 10:48:02.511660 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-ckqhf" event={"ID":"048fff94-2a2e-4819-af68-f3e8b2f2fa3f","Type":"ContainerStarted","Data":"ad328b54ff67c55cd0d3600a0b3e878bcf0e84f743f7363ad5dbf876f2e8e8c9"} Dec 06 10:48:02 crc kubenswrapper[4678]: I1206 10:48:02.514325 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" event={"ID":"882b9c0d-2380-42b7-b13f-480a4f53a258","Type":"ContainerStarted","Data":"9fab12a68890e8677b351d276c37fae7605b60d967a64e73d702b381f9f1163d"} Dec 06 10:48:02 crc kubenswrapper[4678]: I1206 10:48:02.515418 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" event={"ID":"5e3d90b8-1cbb-4329-8d55-911e6dad3243","Type":"ContainerStarted","Data":"57d475f4e028c6caefafc01c6036a2de7bb4330ce6f92e0f15e1315291825b9b"} Dec 06 10:48:05 crc kubenswrapper[4678]: I1206 10:48:05.535451 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" event={"ID":"5e3d90b8-1cbb-4329-8d55-911e6dad3243","Type":"ContainerStarted","Data":"0836cb04dadc4ba944ee08116f81dd79d449753b0359320f95fde856c8b7acb5"} Dec 06 10:48:05 crc kubenswrapper[4678]: I1206 10:48:05.536152 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" Dec 06 10:48:05 crc kubenswrapper[4678]: I1206 10:48:05.537453 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-ckqhf" event={"ID":"048fff94-2a2e-4819-af68-f3e8b2f2fa3f","Type":"ContainerStarted","Data":"944afac187357c9827fbd612300fab36aa5abcc3a10000bec32ffaf5b50b4a27"} Dec 06 10:48:05 crc kubenswrapper[4678]: I1206 10:48:05.538842 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" event={"ID":"882b9c0d-2380-42b7-b13f-480a4f53a258","Type":"ContainerStarted","Data":"390c3b50497dfc3838e605cad8ee8e442d6a5eca2c55ca2d0fb259dab93ac187"} Dec 06 10:48:05 crc kubenswrapper[4678]: I1206 10:48:05.555910 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" podStartSLOduration=1.92289949 podStartE2EDuration="5.555885697s" podCreationTimestamp="2025-12-06 10:48:00 +0000 UTC" firstStartedPulling="2025-12-06 10:48:01.631936164 +0000 UTC m=+686.475367603" lastFinishedPulling="2025-12-06 10:48:05.264922371 +0000 UTC m=+690.108353810" observedRunningTime="2025-12-06 10:48:05.55000205 +0000 UTC m=+690.393433489" watchObservedRunningTime="2025-12-06 10:48:05.555885697 +0000 UTC m=+690.399317136" Dec 06 10:48:05 crc kubenswrapper[4678]: I1206 10:48:05.569746 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-8889z" podStartSLOduration=1.8313932080000002 podStartE2EDuration="5.569713913s" podCreationTimestamp="2025-12-06 10:48:00 +0000 UTC" firstStartedPulling="2025-12-06 10:48:01.519359399 +0000 UTC m=+686.362790838" lastFinishedPulling="2025-12-06 10:48:05.257680114 +0000 UTC m=+690.101111543" observedRunningTime="2025-12-06 10:48:05.565991322 +0000 UTC m=+690.409422761" watchObservedRunningTime="2025-12-06 10:48:05.569713913 +0000 UTC m=+690.413145352" Dec 06 10:48:05 crc kubenswrapper[4678]: I1206 10:48:05.588162 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-ckqhf" podStartSLOduration=1.87933087 podStartE2EDuration="5.588136487s" podCreationTimestamp="2025-12-06 10:48:00 +0000 UTC" firstStartedPulling="2025-12-06 10:48:01.569127245 +0000 UTC m=+686.412558684" lastFinishedPulling="2025-12-06 10:48:05.277932872 +0000 UTC m=+690.121364301" observedRunningTime="2025-12-06 10:48:05.583541988 +0000 UTC m=+690.426973427" watchObservedRunningTime="2025-12-06 10:48:05.588136487 +0000 UTC m=+690.431567946" Dec 06 10:48:11 crc kubenswrapper[4678]: I1206 10:48:11.264318 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-rhwnp" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.119544 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5lfqt"] Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.120446 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-controller" containerID="cri-o://e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.120560 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="northd" containerID="cri-o://7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.120579 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-acl-logging" containerID="cri-o://971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.120605 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-node" containerID="cri-o://0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.120653 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="sbdb" containerID="cri-o://a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.121057 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.120568 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="nbdb" containerID="cri-o://d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.165338 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" containerID="cri-o://b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32" gracePeriod=30 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.614169 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/2.log" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.616168 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/1.log" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.616281 4678 generic.go:334] "Generic (PLEG): container finished" podID="388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c" containerID="1f99ae1ba562ece348c05726186b8fa5e06897de886758153dfe02b1c5a8e84c" exitCode=2 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.616426 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerDied","Data":"1f99ae1ba562ece348c05726186b8fa5e06897de886758153dfe02b1c5a8e84c"} Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.616546 4678 scope.go:117] "RemoveContainer" containerID="278291e25b95f0ed0d93ddff1d4c9a3ea7a50877eed4b547da1444539f7a1e13" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.617353 4678 scope.go:117] "RemoveContainer" containerID="1f99ae1ba562ece348c05726186b8fa5e06897de886758153dfe02b1c5a8e84c" Dec 06 10:48:16 crc kubenswrapper[4678]: E1206 10:48:16.617802 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-z5s25_openshift-multus(388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c)\"" pod="openshift-multus/multus-z5s25" podUID="388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.632532 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovnkube-controller/3.log" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.640364 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovn-acl-logging/0.log" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.642110 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovn-controller/0.log" Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643209 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32" exitCode=0 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643247 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880" exitCode=0 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643257 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a" exitCode=0 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643269 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30" exitCode=143 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643282 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1" exitCode=143 Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643310 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32"} Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643342 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880"} Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.643874 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a"} Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.649733 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30"} Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.649779 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1"} Dec 06 10:48:16 crc kubenswrapper[4678]: I1206 10:48:16.696806 4678 scope.go:117] "RemoveContainer" containerID="226ab684b192938eb482fcbf2e96c6f646252285a17d024763bdb2c10ba207f7" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.168093 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovn-acl-logging/0.log" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.170415 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovn-controller/0.log" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.171171 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.261578 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-gmkg6"] Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262508 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262544 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262575 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262599 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262643 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262654 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262679 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kubecfg-setup" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262689 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kubecfg-setup" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262705 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="northd" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262714 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="northd" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262723 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="sbdb" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262746 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="sbdb" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262775 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262785 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262810 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-acl-logging" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262819 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-acl-logging" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262847 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262871 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262896 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-node" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262906 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-node" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262934 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262945 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.262969 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="nbdb" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.262980 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="nbdb" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263179 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-acl-logging" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263204 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="nbdb" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263216 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263241 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263263 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263316 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263341 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263365 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovn-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263388 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="northd" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263395 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="sbdb" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263428 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-ovn-metrics" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263440 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="kube-rbac-proxy-node" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.263621 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.263638 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerName="ovnkube-controller" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.267838 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315403 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-node-log\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315464 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-env-overrides\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315505 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-ovn\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315540 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-log-socket\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315557 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-systemd-units\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315585 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-bin\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315599 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-slash\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315611 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-var-lib-openvswitch\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315577 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-node-log" (OuterVolumeSpecName: "node-log") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315606 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315650 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-config\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315657 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315665 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-etc-openvswitch\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315680 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315689 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-netns\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315705 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-slash" (OuterVolumeSpecName: "host-slash") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315709 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjjj4\" (UniqueName: \"kubernetes.io/projected/c5cca846-f84b-4924-b1c6-4ec6cea71a65-kube-api-access-qjjj4\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315684 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-log-socket" (OuterVolumeSpecName: "log-socket") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315747 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315777 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-netd\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315784 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315802 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovn-node-metrics-cert\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315824 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-script-lib\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315848 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-ovn-kubernetes\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315878 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-kubelet\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315895 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-systemd\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315894 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315935 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315915 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-openvswitch\") pod \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\" (UID: \"c5cca846-f84b-4924-b1c6-4ec6cea71a65\") " Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315960 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.315984 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316271 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316305 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316435 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316451 4678 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316462 4678 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316475 4678 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316485 4678 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316510 4678 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316520 4678 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-node-log\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316529 4678 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316538 4678 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-log-socket\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316546 4678 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316555 4678 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316565 4678 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-slash\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316575 4678 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316603 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316616 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.316644 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.317227 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.329196 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5cca846-f84b-4924-b1c6-4ec6cea71a65-kube-api-access-qjjj4" (OuterVolumeSpecName: "kube-api-access-qjjj4") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "kube-api-access-qjjj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.329665 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.329747 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c5cca846-f84b-4924-b1c6-4ec6cea71a65" (UID: "c5cca846-f84b-4924-b1c6-4ec6cea71a65"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.417732 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-cni-bin\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418163 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418657 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-env-overrides\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418730 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-kubelet\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418752 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-node-log\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418768 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-run-netns\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418785 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-cni-netd\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418803 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418818 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-ovnkube-config\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418836 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-run-ovn-kubernetes\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418851 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-slash\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418866 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e36370e1-312e-4d30-bef0-89d34f47c59c-ovn-node-metrics-cert\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418892 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-var-lib-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418908 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-ovnkube-script-lib\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418931 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-etc-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.418965 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-systemd\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419010 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-systemd-units\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419026 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-ovn\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419041 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-log-socket\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419079 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx2nq\" (UniqueName: \"kubernetes.io/projected/e36370e1-312e-4d30-bef0-89d34f47c59c-kube-api-access-gx2nq\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419119 4678 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419174 4678 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419444 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjjj4\" (UniqueName: \"kubernetes.io/projected/c5cca846-f84b-4924-b1c6-4ec6cea71a65-kube-api-access-qjjj4\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419539 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419552 4678 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c5cca846-f84b-4924-b1c6-4ec6cea71a65-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419561 4678 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.419568 4678 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c5cca846-f84b-4924-b1c6-4ec6cea71a65-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520430 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-cni-bin\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520606 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-cni-bin\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520622 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520722 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-env-overrides\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520801 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-kubelet\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520833 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-node-log\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520858 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-run-netns\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520874 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-cni-netd\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520878 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-kubelet\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520894 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520942 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-cni-netd\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520946 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-run-netns\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520922 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520950 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-ovnkube-config\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521014 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-run-ovn-kubernetes\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521035 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e36370e1-312e-4d30-bef0-89d34f47c59c-ovn-node-metrics-cert\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521056 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-slash\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521104 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-var-lib-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521134 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-ovnkube-script-lib\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521178 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-etc-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521209 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-var-lib-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521242 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-slash\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.520916 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-node-log\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521268 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-host-run-ovn-kubernetes\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521280 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-systemd\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521337 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-systemd-units\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521362 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-ovn\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521390 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-log-socket\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521411 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx2nq\" (UniqueName: \"kubernetes.io/projected/e36370e1-312e-4d30-bef0-89d34f47c59c-kube-api-access-gx2nq\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521564 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-env-overrides\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521639 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-ovnkube-config\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521879 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-etc-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521905 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-ovn\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521933 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-log-socket\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521937 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-systemd-units\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.521968 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-systemd\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.522113 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e36370e1-312e-4d30-bef0-89d34f47c59c-run-openvswitch\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.522212 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e36370e1-312e-4d30-bef0-89d34f47c59c-ovnkube-script-lib\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.525976 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e36370e1-312e-4d30-bef0-89d34f47c59c-ovn-node-metrics-cert\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.538199 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx2nq\" (UniqueName: \"kubernetes.io/projected/e36370e1-312e-4d30-bef0-89d34f47c59c-kube-api-access-gx2nq\") pod \"ovnkube-node-gmkg6\" (UID: \"e36370e1-312e-4d30-bef0-89d34f47c59c\") " pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.587446 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.650745 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/2.log" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.655384 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovn-acl-logging/0.log" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.655926 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-5lfqt_c5cca846-f84b-4924-b1c6-4ec6cea71a65/ovn-controller/0.log" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656367 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9" exitCode=0 Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656433 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5" exitCode=0 Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656448 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" containerID="7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e" exitCode=0 Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656475 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9"} Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656553 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5"} Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656571 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e"} Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656579 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656598 4678 scope.go:117] "RemoveContainer" containerID="b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.656583 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5lfqt" event={"ID":"c5cca846-f84b-4924-b1c6-4ec6cea71a65","Type":"ContainerDied","Data":"96ba415b0bee77024b2bb13f377a61308cd4a59a064d968e1201ee85e1c014df"} Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.657920 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"22291498be644f3935acbff71f032d318846cb173a762550689b31908ad85dba"} Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.683604 4678 scope.go:117] "RemoveContainer" containerID="a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.690202 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5lfqt"] Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.699084 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5lfqt"] Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.701833 4678 scope.go:117] "RemoveContainer" containerID="d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.718330 4678 scope.go:117] "RemoveContainer" containerID="7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.740131 4678 scope.go:117] "RemoveContainer" containerID="e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.781824 4678 scope.go:117] "RemoveContainer" containerID="0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.795380 4678 scope.go:117] "RemoveContainer" containerID="971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.808215 4678 scope.go:117] "RemoveContainer" containerID="e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.822483 4678 scope.go:117] "RemoveContainer" containerID="01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.836839 4678 scope.go:117] "RemoveContainer" containerID="b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.837261 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32\": container with ID starting with b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32 not found: ID does not exist" containerID="b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.837325 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32"} err="failed to get container status \"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32\": rpc error: code = NotFound desc = could not find container \"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32\": container with ID starting with b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.837363 4678 scope.go:117] "RemoveContainer" containerID="a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.837709 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\": container with ID starting with a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9 not found: ID does not exist" containerID="a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.837739 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9"} err="failed to get container status \"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\": rpc error: code = NotFound desc = could not find container \"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\": container with ID starting with a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.837754 4678 scope.go:117] "RemoveContainer" containerID="d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.838014 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\": container with ID starting with d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5 not found: ID does not exist" containerID="d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.838040 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5"} err="failed to get container status \"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\": rpc error: code = NotFound desc = could not find container \"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\": container with ID starting with d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.838056 4678 scope.go:117] "RemoveContainer" containerID="7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.838341 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\": container with ID starting with 7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e not found: ID does not exist" containerID="7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.838371 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e"} err="failed to get container status \"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\": rpc error: code = NotFound desc = could not find container \"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\": container with ID starting with 7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.838387 4678 scope.go:117] "RemoveContainer" containerID="e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.838769 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\": container with ID starting with e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880 not found: ID does not exist" containerID="e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.838797 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880"} err="failed to get container status \"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\": rpc error: code = NotFound desc = could not find container \"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\": container with ID starting with e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.838819 4678 scope.go:117] "RemoveContainer" containerID="0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.839332 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\": container with ID starting with 0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a not found: ID does not exist" containerID="0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.839362 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a"} err="failed to get container status \"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\": rpc error: code = NotFound desc = could not find container \"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\": container with ID starting with 0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.839418 4678 scope.go:117] "RemoveContainer" containerID="971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.839686 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\": container with ID starting with 971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30 not found: ID does not exist" containerID="971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.839710 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30"} err="failed to get container status \"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\": rpc error: code = NotFound desc = could not find container \"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\": container with ID starting with 971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.839726 4678 scope.go:117] "RemoveContainer" containerID="e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.840012 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\": container with ID starting with e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1 not found: ID does not exist" containerID="e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.840034 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1"} err="failed to get container status \"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\": rpc error: code = NotFound desc = could not find container \"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\": container with ID starting with e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.840049 4678 scope.go:117] "RemoveContainer" containerID="01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5" Dec 06 10:48:17 crc kubenswrapper[4678]: E1206 10:48:17.840258 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\": container with ID starting with 01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5 not found: ID does not exist" containerID="01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.840301 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5"} err="failed to get container status \"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\": rpc error: code = NotFound desc = could not find container \"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\": container with ID starting with 01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.840314 4678 scope.go:117] "RemoveContainer" containerID="b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.840775 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32"} err="failed to get container status \"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32\": rpc error: code = NotFound desc = could not find container \"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32\": container with ID starting with b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.840793 4678 scope.go:117] "RemoveContainer" containerID="a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841059 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9"} err="failed to get container status \"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\": rpc error: code = NotFound desc = could not find container \"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\": container with ID starting with a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841074 4678 scope.go:117] "RemoveContainer" containerID="d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841322 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5"} err="failed to get container status \"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\": rpc error: code = NotFound desc = could not find container \"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\": container with ID starting with d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841334 4678 scope.go:117] "RemoveContainer" containerID="7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841591 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e"} err="failed to get container status \"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\": rpc error: code = NotFound desc = could not find container \"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\": container with ID starting with 7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841615 4678 scope.go:117] "RemoveContainer" containerID="e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841893 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880"} err="failed to get container status \"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\": rpc error: code = NotFound desc = could not find container \"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\": container with ID starting with e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.841912 4678 scope.go:117] "RemoveContainer" containerID="0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.842134 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a"} err="failed to get container status \"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\": rpc error: code = NotFound desc = could not find container \"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\": container with ID starting with 0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.842157 4678 scope.go:117] "RemoveContainer" containerID="971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.842355 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30"} err="failed to get container status \"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\": rpc error: code = NotFound desc = could not find container \"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\": container with ID starting with 971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.842372 4678 scope.go:117] "RemoveContainer" containerID="e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.842715 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1"} err="failed to get container status \"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\": rpc error: code = NotFound desc = could not find container \"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\": container with ID starting with e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.842760 4678 scope.go:117] "RemoveContainer" containerID="01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843008 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5"} err="failed to get container status \"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\": rpc error: code = NotFound desc = could not find container \"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\": container with ID starting with 01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843022 4678 scope.go:117] "RemoveContainer" containerID="b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843222 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32"} err="failed to get container status \"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32\": rpc error: code = NotFound desc = could not find container \"b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32\": container with ID starting with b4fbc5c595755fb9b0e637effd8112ff2bdfa2a2476d76f2fbe71f64b0561e32 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843236 4678 scope.go:117] "RemoveContainer" containerID="a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843420 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9"} err="failed to get container status \"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\": rpc error: code = NotFound desc = could not find container \"a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9\": container with ID starting with a8e1b25c1595022312c3f4f25bc74427f18096f38b4da9673dda01ffed2bdba9 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843436 4678 scope.go:117] "RemoveContainer" containerID="d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843766 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5"} err="failed to get container status \"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\": rpc error: code = NotFound desc = could not find container \"d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5\": container with ID starting with d94bad61b8280957fb6ce561bd811eff9cf60d5a230ee89c06fa2b4aabbea4f5 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843786 4678 scope.go:117] "RemoveContainer" containerID="7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.843995 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e"} err="failed to get container status \"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\": rpc error: code = NotFound desc = could not find container \"7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e\": container with ID starting with 7170ee1a9af41133d573662985090ef01b61229b79dfdb2c6cc29cba6cf4d96e not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.844014 4678 scope.go:117] "RemoveContainer" containerID="e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.844239 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880"} err="failed to get container status \"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\": rpc error: code = NotFound desc = could not find container \"e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880\": container with ID starting with e495d865963f93a25f115edcf60f7f8fb7f9be0276e2c4a95004bd8b9fe6b880 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.844258 4678 scope.go:117] "RemoveContainer" containerID="0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.844544 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a"} err="failed to get container status \"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\": rpc error: code = NotFound desc = could not find container \"0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a\": container with ID starting with 0e3e71aca1fdbcabacb8b32457a4562fd16ec5dcf35a3d3913e2cd29babd885a not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.844564 4678 scope.go:117] "RemoveContainer" containerID="971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.844813 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30"} err="failed to get container status \"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\": rpc error: code = NotFound desc = could not find container \"971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30\": container with ID starting with 971406ac3cc6a4b6e398fd763b786a0b9048bb89506709be87cd084e4f4dee30 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.844827 4678 scope.go:117] "RemoveContainer" containerID="e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.845067 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1"} err="failed to get container status \"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\": rpc error: code = NotFound desc = could not find container \"e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1\": container with ID starting with e02f45b331eadebb135f6b30b9797a34b277c7315f8cf9377a2d4a8a78460de1 not found: ID does not exist" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.845079 4678 scope.go:117] "RemoveContainer" containerID="01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5" Dec 06 10:48:17 crc kubenswrapper[4678]: I1206 10:48:17.845258 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5"} err="failed to get container status \"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\": rpc error: code = NotFound desc = could not find container \"01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5\": container with ID starting with 01cbea6140ee465ac3ef5c347f81658561a44411508622db80b1ebb6e401aca5 not found: ID does not exist" Dec 06 10:48:18 crc kubenswrapper[4678]: I1206 10:48:18.702404 4678 generic.go:334] "Generic (PLEG): container finished" podID="e36370e1-312e-4d30-bef0-89d34f47c59c" containerID="f77710cc0baa0e22b363ab8257b556e46876591bb2656ff78a557bc1d34ef6ab" exitCode=0 Dec 06 10:48:18 crc kubenswrapper[4678]: I1206 10:48:18.702451 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerDied","Data":"f77710cc0baa0e22b363ab8257b556e46876591bb2656ff78a557bc1d34ef6ab"} Dec 06 10:48:19 crc kubenswrapper[4678]: I1206 10:48:19.486964 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5cca846-f84b-4924-b1c6-4ec6cea71a65" path="/var/lib/kubelet/pods/c5cca846-f84b-4924-b1c6-4ec6cea71a65/volumes" Dec 06 10:48:19 crc kubenswrapper[4678]: I1206 10:48:19.713188 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"814e7b6ac3d507a1e648f84f12683dd05e5db968b3b11bb78a3583796031985a"} Dec 06 10:48:19 crc kubenswrapper[4678]: I1206 10:48:19.713247 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"0fe083e01364488c0b8a9d4dbedc530e98a1af0b7ad5c7fa9de52676d9550d43"} Dec 06 10:48:19 crc kubenswrapper[4678]: I1206 10:48:19.713260 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"38265f3de5c7a6dd2ca1559388d694d85037a67bb0ee4fce9edeb48bdbf2f4de"} Dec 06 10:48:19 crc kubenswrapper[4678]: I1206 10:48:19.713269 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"77c7d3cad955a6afcb3d4c0ff695fa5f1a859990f6fc0481671e3fc2780303b7"} Dec 06 10:48:19 crc kubenswrapper[4678]: I1206 10:48:19.713277 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"d3129344d91295d76bd6ec3413e185e5ebc9c1f0f92c79805bb3bb5033a93491"} Dec 06 10:48:19 crc kubenswrapper[4678]: I1206 10:48:19.713288 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"5cb6696bea0ef3186b05e9df616cb6e66ab0bd9638112015a23f56be7e4f5652"} Dec 06 10:48:21 crc kubenswrapper[4678]: I1206 10:48:21.734412 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"c6fe9ee02ff97c8fbe7b7e03a05eea1d754168b37be17bed9b10196110d2da4e"} Dec 06 10:48:24 crc kubenswrapper[4678]: I1206 10:48:24.762073 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" event={"ID":"e36370e1-312e-4d30-bef0-89d34f47c59c","Type":"ContainerStarted","Data":"a959d3e63ba7473b2835307e8fdb8fbbecdb6b696f0b5da3395b4729fe6e6c8f"} Dec 06 10:48:24 crc kubenswrapper[4678]: I1206 10:48:24.762826 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:24 crc kubenswrapper[4678]: I1206 10:48:24.799550 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" podStartSLOduration=7.799520335 podStartE2EDuration="7.799520335s" podCreationTimestamp="2025-12-06 10:48:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:48:24.788986668 +0000 UTC m=+709.632418147" watchObservedRunningTime="2025-12-06 10:48:24.799520335 +0000 UTC m=+709.642951814" Dec 06 10:48:24 crc kubenswrapper[4678]: I1206 10:48:24.811678 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:25 crc kubenswrapper[4678]: I1206 10:48:25.770262 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:25 crc kubenswrapper[4678]: I1206 10:48:25.770333 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:25 crc kubenswrapper[4678]: I1206 10:48:25.812435 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:29 crc kubenswrapper[4678]: I1206 10:48:29.506110 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:48:29 crc kubenswrapper[4678]: I1206 10:48:29.507179 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:48:31 crc kubenswrapper[4678]: I1206 10:48:31.476090 4678 scope.go:117] "RemoveContainer" containerID="1f99ae1ba562ece348c05726186b8fa5e06897de886758153dfe02b1c5a8e84c" Dec 06 10:48:31 crc kubenswrapper[4678]: E1206 10:48:31.476350 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-z5s25_openshift-multus(388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c)\"" pod="openshift-multus/multus-z5s25" podUID="388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c" Dec 06 10:48:42 crc kubenswrapper[4678]: I1206 10:48:42.475990 4678 scope.go:117] "RemoveContainer" containerID="1f99ae1ba562ece348c05726186b8fa5e06897de886758153dfe02b1c5a8e84c" Dec 06 10:48:42 crc kubenswrapper[4678]: I1206 10:48:42.894021 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-z5s25_388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c/kube-multus/2.log" Dec 06 10:48:42 crc kubenswrapper[4678]: I1206 10:48:42.894532 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-z5s25" event={"ID":"388c6b32-5008-4e09-8ed4-2dc4e3b1ac1c","Type":"ContainerStarted","Data":"5cd8cd25a7134b4f7524d3aabaac8ec7bf880614328298f5c426b1408be15867"} Dec 06 10:48:47 crc kubenswrapper[4678]: I1206 10:48:47.637134 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-gmkg6" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.534562 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4"] Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.536169 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.538056 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.552549 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4"] Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.653128 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.653206 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86zgz\" (UniqueName: \"kubernetes.io/projected/53e84367-27a1-4944-9e9b-bc9743e57398-kube-api-access-86zgz\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.653248 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.754457 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86zgz\" (UniqueName: \"kubernetes.io/projected/53e84367-27a1-4944-9e9b-bc9743e57398-kube-api-access-86zgz\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.754601 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.754709 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.755186 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.755327 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.779217 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86zgz\" (UniqueName: \"kubernetes.io/projected/53e84367-27a1-4944-9e9b-bc9743e57398-kube-api-access-86zgz\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:52 crc kubenswrapper[4678]: I1206 10:48:52.854229 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:53 crc kubenswrapper[4678]: I1206 10:48:53.098079 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4"] Dec 06 10:48:53 crc kubenswrapper[4678]: I1206 10:48:53.967268 4678 generic.go:334] "Generic (PLEG): container finished" podID="53e84367-27a1-4944-9e9b-bc9743e57398" containerID="86de9e1f0ecae302307493d809ebb6defba7239579ace52f7acd86d9b8443cad" exitCode=0 Dec 06 10:48:53 crc kubenswrapper[4678]: I1206 10:48:53.967337 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" event={"ID":"53e84367-27a1-4944-9e9b-bc9743e57398","Type":"ContainerDied","Data":"86de9e1f0ecae302307493d809ebb6defba7239579ace52f7acd86d9b8443cad"} Dec 06 10:48:53 crc kubenswrapper[4678]: I1206 10:48:53.967386 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" event={"ID":"53e84367-27a1-4944-9e9b-bc9743e57398","Type":"ContainerStarted","Data":"be38584a61f3c6f06f991901b65af001b6608fcd1a680b6f89ce306a37c2e320"} Dec 06 10:48:55 crc kubenswrapper[4678]: I1206 10:48:55.983708 4678 generic.go:334] "Generic (PLEG): container finished" podID="53e84367-27a1-4944-9e9b-bc9743e57398" containerID="2a9a0aa444bd2c31ec0002070abc378b1d3c7258d4a9a5b5afff3f4f7daa3963" exitCode=0 Dec 06 10:48:55 crc kubenswrapper[4678]: I1206 10:48:55.983935 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" event={"ID":"53e84367-27a1-4944-9e9b-bc9743e57398","Type":"ContainerDied","Data":"2a9a0aa444bd2c31ec0002070abc378b1d3c7258d4a9a5b5afff3f4f7daa3963"} Dec 06 10:48:56 crc kubenswrapper[4678]: I1206 10:48:56.991660 4678 generic.go:334] "Generic (PLEG): container finished" podID="53e84367-27a1-4944-9e9b-bc9743e57398" containerID="d9df9d00795c10e045d535b7adb8be3f27e012b88e08791bbda7fad5958012f3" exitCode=0 Dec 06 10:48:56 crc kubenswrapper[4678]: I1206 10:48:56.991711 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" event={"ID":"53e84367-27a1-4944-9e9b-bc9743e57398","Type":"ContainerDied","Data":"d9df9d00795c10e045d535b7adb8be3f27e012b88e08791bbda7fad5958012f3"} Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.287779 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.437817 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-util\") pod \"53e84367-27a1-4944-9e9b-bc9743e57398\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.437964 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86zgz\" (UniqueName: \"kubernetes.io/projected/53e84367-27a1-4944-9e9b-bc9743e57398-kube-api-access-86zgz\") pod \"53e84367-27a1-4944-9e9b-bc9743e57398\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.438812 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-bundle\") pod \"53e84367-27a1-4944-9e9b-bc9743e57398\" (UID: \"53e84367-27a1-4944-9e9b-bc9743e57398\") " Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.439426 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-bundle" (OuterVolumeSpecName: "bundle") pod "53e84367-27a1-4944-9e9b-bc9743e57398" (UID: "53e84367-27a1-4944-9e9b-bc9743e57398"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.439624 4678 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.447791 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53e84367-27a1-4944-9e9b-bc9743e57398-kube-api-access-86zgz" (OuterVolumeSpecName: "kube-api-access-86zgz") pod "53e84367-27a1-4944-9e9b-bc9743e57398" (UID: "53e84367-27a1-4944-9e9b-bc9743e57398"). InnerVolumeSpecName "kube-api-access-86zgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.452152 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-util" (OuterVolumeSpecName: "util") pod "53e84367-27a1-4944-9e9b-bc9743e57398" (UID: "53e84367-27a1-4944-9e9b-bc9743e57398"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.540909 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86zgz\" (UniqueName: \"kubernetes.io/projected/53e84367-27a1-4944-9e9b-bc9743e57398-kube-api-access-86zgz\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:58 crc kubenswrapper[4678]: I1206 10:48:58.540950 4678 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/53e84367-27a1-4944-9e9b-bc9743e57398-util\") on node \"crc\" DevicePath \"\"" Dec 06 10:48:59 crc kubenswrapper[4678]: I1206 10:48:59.009679 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" event={"ID":"53e84367-27a1-4944-9e9b-bc9743e57398","Type":"ContainerDied","Data":"be38584a61f3c6f06f991901b65af001b6608fcd1a680b6f89ce306a37c2e320"} Dec 06 10:48:59 crc kubenswrapper[4678]: I1206 10:48:59.010300 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be38584a61f3c6f06f991901b65af001b6608fcd1a680b6f89ce306a37c2e320" Dec 06 10:48:59 crc kubenswrapper[4678]: I1206 10:48:59.009801 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4" Dec 06 10:48:59 crc kubenswrapper[4678]: I1206 10:48:59.506315 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:48:59 crc kubenswrapper[4678]: I1206 10:48:59.506436 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.059196 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh"] Dec 06 10:49:00 crc kubenswrapper[4678]: E1206 10:49:00.059416 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e84367-27a1-4944-9e9b-bc9743e57398" containerName="extract" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.059429 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e84367-27a1-4944-9e9b-bc9743e57398" containerName="extract" Dec 06 10:49:00 crc kubenswrapper[4678]: E1206 10:49:00.059454 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e84367-27a1-4944-9e9b-bc9743e57398" containerName="util" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.059459 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e84367-27a1-4944-9e9b-bc9743e57398" containerName="util" Dec 06 10:49:00 crc kubenswrapper[4678]: E1206 10:49:00.059472 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53e84367-27a1-4944-9e9b-bc9743e57398" containerName="pull" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.059522 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="53e84367-27a1-4944-9e9b-bc9743e57398" containerName="pull" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.059611 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="53e84367-27a1-4944-9e9b-bc9743e57398" containerName="extract" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.059978 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.063709 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.064054 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.067099 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-8trq5" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.080785 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh"] Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.162858 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khrwt\" (UniqueName: \"kubernetes.io/projected/fd2552d3-a5fd-4b13-8d01-996a25348826-kube-api-access-khrwt\") pod \"nmstate-operator-5b5b58f5c8-5rqmh\" (UID: \"fd2552d3-a5fd-4b13-8d01-996a25348826\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.264652 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khrwt\" (UniqueName: \"kubernetes.io/projected/fd2552d3-a5fd-4b13-8d01-996a25348826-kube-api-access-khrwt\") pod \"nmstate-operator-5b5b58f5c8-5rqmh\" (UID: \"fd2552d3-a5fd-4b13-8d01-996a25348826\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.281905 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khrwt\" (UniqueName: \"kubernetes.io/projected/fd2552d3-a5fd-4b13-8d01-996a25348826-kube-api-access-khrwt\") pod \"nmstate-operator-5b5b58f5c8-5rqmh\" (UID: \"fd2552d3-a5fd-4b13-8d01-996a25348826\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.382705 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" Dec 06 10:49:00 crc kubenswrapper[4678]: I1206 10:49:00.582250 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh"] Dec 06 10:49:01 crc kubenswrapper[4678]: I1206 10:49:01.021788 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" event={"ID":"fd2552d3-a5fd-4b13-8d01-996a25348826","Type":"ContainerStarted","Data":"f2eedbe71ffa90b05b6952ac2b184e98001b71cf0cd484ac0b16bfcd95e4931a"} Dec 06 10:49:04 crc kubenswrapper[4678]: I1206 10:49:04.045938 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" event={"ID":"fd2552d3-a5fd-4b13-8d01-996a25348826","Type":"ContainerStarted","Data":"30cd693829167cea811fa8db0e6dedce7615dc7c738352c6dd3036676fc08411"} Dec 06 10:49:04 crc kubenswrapper[4678]: I1206 10:49:04.078176 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5rqmh" podStartSLOduration=1.342055676 podStartE2EDuration="4.078151634s" podCreationTimestamp="2025-12-06 10:49:00 +0000 UTC" firstStartedPulling="2025-12-06 10:49:00.607200659 +0000 UTC m=+745.450632088" lastFinishedPulling="2025-12-06 10:49:03.343296607 +0000 UTC m=+748.186728046" observedRunningTime="2025-12-06 10:49:04.07095083 +0000 UTC m=+748.914382279" watchObservedRunningTime="2025-12-06 10:49:04.078151634 +0000 UTC m=+748.921583083" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.019253 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.020368 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.029932 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.032282 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.043855 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-j2bt9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.044126 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.068610 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.101591 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.124434 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-n2j4m"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.125293 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.136141 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpsq6\" (UniqueName: \"kubernetes.io/projected/ca2c68b1-3733-4be8-9580-616e9d436a34-kube-api-access-jpsq6\") pod \"nmstate-webhook-5f6d4c5ccb-8gwjz\" (UID: \"ca2c68b1-3733-4be8-9580-616e9d436a34\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.136570 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85xc4\" (UniqueName: \"kubernetes.io/projected/1e102849-cdd0-4035-a9ba-339d77bd6f8e-kube-api-access-85xc4\") pod \"nmstate-metrics-7f946cbc9-7nc8k\" (UID: \"1e102849-cdd0-4035-a9ba-339d77bd6f8e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.136618 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ca2c68b1-3733-4be8-9580-616e9d436a34-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8gwjz\" (UID: \"ca2c68b1-3733-4be8-9580-616e9d436a34\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.220825 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.221589 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.227930 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.228674 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.228789 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-wlv6c" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.237338 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-ovs-socket\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.237396 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44pwm\" (UniqueName: \"kubernetes.io/projected/82398c4a-8182-4723-bad7-0c289be9d283-kube-api-access-44pwm\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.237425 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpsq6\" (UniqueName: \"kubernetes.io/projected/ca2c68b1-3733-4be8-9580-616e9d436a34-kube-api-access-jpsq6\") pod \"nmstate-webhook-5f6d4c5ccb-8gwjz\" (UID: \"ca2c68b1-3733-4be8-9580-616e9d436a34\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.237443 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85xc4\" (UniqueName: \"kubernetes.io/projected/1e102849-cdd0-4035-a9ba-339d77bd6f8e-kube-api-access-85xc4\") pod \"nmstate-metrics-7f946cbc9-7nc8k\" (UID: \"1e102849-cdd0-4035-a9ba-339d77bd6f8e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.237467 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-dbus-socket\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.237507 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ca2c68b1-3733-4be8-9580-616e9d436a34-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8gwjz\" (UID: \"ca2c68b1-3733-4be8-9580-616e9d436a34\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.237559 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-nmstate-lock\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: E1206 10:49:05.238080 4678 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 06 10:49:05 crc kubenswrapper[4678]: E1206 10:49:05.238132 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ca2c68b1-3733-4be8-9580-616e9d436a34-tls-key-pair podName:ca2c68b1-3733-4be8-9580-616e9d436a34 nodeName:}" failed. No retries permitted until 2025-12-06 10:49:05.738111559 +0000 UTC m=+750.581542988 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ca2c68b1-3733-4be8-9580-616e9d436a34-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-8gwjz" (UID: "ca2c68b1-3733-4be8-9580-616e9d436a34") : secret "openshift-nmstate-webhook" not found Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.242824 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.265419 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpsq6\" (UniqueName: \"kubernetes.io/projected/ca2c68b1-3733-4be8-9580-616e9d436a34-kube-api-access-jpsq6\") pod \"nmstate-webhook-5f6d4c5ccb-8gwjz\" (UID: \"ca2c68b1-3733-4be8-9580-616e9d436a34\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.278446 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85xc4\" (UniqueName: \"kubernetes.io/projected/1e102849-cdd0-4035-a9ba-339d77bd6f8e-kube-api-access-85xc4\") pod \"nmstate-metrics-7f946cbc9-7nc8k\" (UID: \"1e102849-cdd0-4035-a9ba-339d77bd6f8e\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.338976 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-ovs-socket\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339315 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44pwm\" (UniqueName: \"kubernetes.io/projected/82398c4a-8182-4723-bad7-0c289be9d283-kube-api-access-44pwm\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339422 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d99e972f-a316-4107-85b4-bf5411999e91-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339525 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-dbus-socket\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339075 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-ovs-socket\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339719 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-nmstate-lock\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339866 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ngw4\" (UniqueName: \"kubernetes.io/projected/d99e972f-a316-4107-85b4-bf5411999e91-kube-api-access-6ngw4\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339764 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-nmstate-lock\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.339939 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d99e972f-a316-4107-85b4-bf5411999e91-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.340064 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/82398c4a-8182-4723-bad7-0c289be9d283-dbus-socket\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.367762 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44pwm\" (UniqueName: \"kubernetes.io/projected/82398c4a-8182-4723-bad7-0c289be9d283-kube-api-access-44pwm\") pod \"nmstate-handler-n2j4m\" (UID: \"82398c4a-8182-4723-bad7-0c289be9d283\") " pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.371156 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.432394 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-c76489864-zvgk9"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.433370 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.443659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ngw4\" (UniqueName: \"kubernetes.io/projected/d99e972f-a316-4107-85b4-bf5411999e91-kube-api-access-6ngw4\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.443715 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d99e972f-a316-4107-85b4-bf5411999e91-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.443766 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d99e972f-a316-4107-85b4-bf5411999e91-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.445417 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d99e972f-a316-4107-85b4-bf5411999e91-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.451200 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d99e972f-a316-4107-85b4-bf5411999e91-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.451438 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.461909 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-c76489864-zvgk9"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.491470 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ngw4\" (UniqueName: \"kubernetes.io/projected/d99e972f-a316-4107-85b4-bf5411999e91-kube-api-access-6ngw4\") pod \"nmstate-console-plugin-7fbb5f6569-g72ds\" (UID: \"d99e972f-a316-4107-85b4-bf5411999e91\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: W1206 10:49:05.504275 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82398c4a_8182_4723_bad7_0c289be9d283.slice/crio-4b69ed087625608bcc46aad33a92c02003d1f6850147bd8f1964eaf138dae9ed WatchSource:0}: Error finding container 4b69ed087625608bcc46aad33a92c02003d1f6850147bd8f1964eaf138dae9ed: Status 404 returned error can't find the container with id 4b69ed087625608bcc46aad33a92c02003d1f6850147bd8f1964eaf138dae9ed Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.544540 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-serving-cert\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.545053 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-oauth-serving-cert\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.545078 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-oauth-config\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.545111 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-trusted-ca-bundle\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.545135 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-service-ca\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.545224 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q776d\" (UniqueName: \"kubernetes.io/projected/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-kube-api-access-q776d\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.545271 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-config\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.545676 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.646694 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q776d\" (UniqueName: \"kubernetes.io/projected/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-kube-api-access-q776d\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.646756 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-config\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.646807 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-serving-cert\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.646828 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-oauth-serving-cert\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.646850 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-oauth-config\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.646873 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-trusted-ca-bundle\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.646893 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-service-ca\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.647897 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-service-ca\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.648308 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-oauth-serving-cert\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.648898 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-config\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.649623 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-trusted-ca-bundle\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.654988 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-serving-cert\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.665212 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-console-oauth-config\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.670071 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q776d\" (UniqueName: \"kubernetes.io/projected/9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d-kube-api-access-q776d\") pod \"console-c76489864-zvgk9\" (UID: \"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d\") " pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.748420 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ca2c68b1-3733-4be8-9580-616e9d436a34-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8gwjz\" (UID: \"ca2c68b1-3733-4be8-9580-616e9d436a34\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.751880 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ca2c68b1-3733-4be8-9580-616e9d436a34-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-8gwjz\" (UID: \"ca2c68b1-3733-4be8-9580-616e9d436a34\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.785615 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.889885 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k"] Dec 06 10:49:05 crc kubenswrapper[4678]: W1206 10:49:05.901628 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1e102849_cdd0_4035_a9ba_339d77bd6f8e.slice/crio-2539968c8d5ee7d0a1c4b3555a14504d3b9b3a26b06466d4420c0f4be57a1327 WatchSource:0}: Error finding container 2539968c8d5ee7d0a1c4b3555a14504d3b9b3a26b06466d4420c0f4be57a1327: Status 404 returned error can't find the container with id 2539968c8d5ee7d0a1c4b3555a14504d3b9b3a26b06466d4420c0f4be57a1327 Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.981461 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds"] Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.984357 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:05 crc kubenswrapper[4678]: I1206 10:49:05.989804 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-c76489864-zvgk9"] Dec 06 10:49:05 crc kubenswrapper[4678]: W1206 10:49:05.993926 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f2ef5b1_62f4_4bcf_9b87_e17ab4f4ab9d.slice/crio-e27c1b4903a3ff6e85f197037218572378e988a84842d77b5ae45dc79ff1f1c3 WatchSource:0}: Error finding container e27c1b4903a3ff6e85f197037218572378e988a84842d77b5ae45dc79ff1f1c3: Status 404 returned error can't find the container with id e27c1b4903a3ff6e85f197037218572378e988a84842d77b5ae45dc79ff1f1c3 Dec 06 10:49:06 crc kubenswrapper[4678]: I1206 10:49:06.066073 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" event={"ID":"d99e972f-a316-4107-85b4-bf5411999e91","Type":"ContainerStarted","Data":"4e971540a26a809263f74d6157882fedac67db912984f47a2d06360904ddceaf"} Dec 06 10:49:06 crc kubenswrapper[4678]: I1206 10:49:06.067169 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" event={"ID":"1e102849-cdd0-4035-a9ba-339d77bd6f8e","Type":"ContainerStarted","Data":"2539968c8d5ee7d0a1c4b3555a14504d3b9b3a26b06466d4420c0f4be57a1327"} Dec 06 10:49:06 crc kubenswrapper[4678]: I1206 10:49:06.068150 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-n2j4m" event={"ID":"82398c4a-8182-4723-bad7-0c289be9d283","Type":"ContainerStarted","Data":"4b69ed087625608bcc46aad33a92c02003d1f6850147bd8f1964eaf138dae9ed"} Dec 06 10:49:06 crc kubenswrapper[4678]: I1206 10:49:06.068911 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c76489864-zvgk9" event={"ID":"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d","Type":"ContainerStarted","Data":"e27c1b4903a3ff6e85f197037218572378e988a84842d77b5ae45dc79ff1f1c3"} Dec 06 10:49:06 crc kubenswrapper[4678]: I1206 10:49:06.221725 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz"] Dec 06 10:49:06 crc kubenswrapper[4678]: W1206 10:49:06.233567 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca2c68b1_3733_4be8_9580_616e9d436a34.slice/crio-c1ea1e0c95f9be0fe970d0d7c8b22b33b8e77551e27865d38639f95e7a793020 WatchSource:0}: Error finding container c1ea1e0c95f9be0fe970d0d7c8b22b33b8e77551e27865d38639f95e7a793020: Status 404 returned error can't find the container with id c1ea1e0c95f9be0fe970d0d7c8b22b33b8e77551e27865d38639f95e7a793020 Dec 06 10:49:07 crc kubenswrapper[4678]: I1206 10:49:07.076772 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" event={"ID":"ca2c68b1-3733-4be8-9580-616e9d436a34","Type":"ContainerStarted","Data":"c1ea1e0c95f9be0fe970d0d7c8b22b33b8e77551e27865d38639f95e7a793020"} Dec 06 10:49:07 crc kubenswrapper[4678]: I1206 10:49:07.078434 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-c76489864-zvgk9" event={"ID":"9f2ef5b1-62f4-4bcf-9b87-e17ab4f4ab9d","Type":"ContainerStarted","Data":"3dda90301d59adf6e8d6d27df9e423086b31f48004638bcaa9f32e4ea22c8559"} Dec 06 10:49:07 crc kubenswrapper[4678]: I1206 10:49:07.100708 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-c76489864-zvgk9" podStartSLOduration=2.100686364 podStartE2EDuration="2.100686364s" podCreationTimestamp="2025-12-06 10:49:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:49:07.097130554 +0000 UTC m=+751.940561993" watchObservedRunningTime="2025-12-06 10:49:07.100686364 +0000 UTC m=+751.944117803" Dec 06 10:49:08 crc kubenswrapper[4678]: I1206 10:49:08.906233 4678 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.098698 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" event={"ID":"ca2c68b1-3733-4be8-9580-616e9d436a34","Type":"ContainerStarted","Data":"d88dc45fb4ee7dbcd3ad1bd58081bea21789fdd15992196b3016332863a506ce"} Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.099897 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.101040 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-n2j4m" event={"ID":"82398c4a-8182-4723-bad7-0c289be9d283","Type":"ContainerStarted","Data":"51969b133aa3c0be894c2fe77ddff31ece42b57330c46fd62fa9e2ec65ddf358"} Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.101945 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.109618 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" event={"ID":"d99e972f-a316-4107-85b4-bf5411999e91","Type":"ContainerStarted","Data":"5271fc01dd3804d3c474438c97815371e5f8e46126e673c20b868956b54b78c9"} Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.114870 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" event={"ID":"1e102849-cdd0-4035-a9ba-339d77bd6f8e","Type":"ContainerStarted","Data":"b223cf90f6324a75a524713c645b7445528b80a3ac33914856932efe489e2f6e"} Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.127205 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" podStartSLOduration=2.147964029 podStartE2EDuration="5.127180229s" podCreationTimestamp="2025-12-06 10:49:05 +0000 UTC" firstStartedPulling="2025-12-06 10:49:06.236875806 +0000 UTC m=+751.080307245" lastFinishedPulling="2025-12-06 10:49:09.216092006 +0000 UTC m=+754.059523445" observedRunningTime="2025-12-06 10:49:10.126018203 +0000 UTC m=+754.969449672" watchObservedRunningTime="2025-12-06 10:49:10.127180229 +0000 UTC m=+754.970611668" Dec 06 10:49:10 crc kubenswrapper[4678]: I1206 10:49:10.157237 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-g72ds" podStartSLOduration=1.9655941220000002 podStartE2EDuration="5.157218807s" podCreationTimestamp="2025-12-06 10:49:05 +0000 UTC" firstStartedPulling="2025-12-06 10:49:05.988846981 +0000 UTC m=+750.832278420" lastFinishedPulling="2025-12-06 10:49:09.180471666 +0000 UTC m=+754.023903105" observedRunningTime="2025-12-06 10:49:10.155587295 +0000 UTC m=+754.999018734" watchObservedRunningTime="2025-12-06 10:49:10.157218807 +0000 UTC m=+755.000650246" Dec 06 10:49:12 crc kubenswrapper[4678]: I1206 10:49:12.138299 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" event={"ID":"1e102849-cdd0-4035-a9ba-339d77bd6f8e","Type":"ContainerStarted","Data":"d1443dd4a336d609e5ad1b3294557642de9802fe77ea10f649c019523e144c79"} Dec 06 10:49:12 crc kubenswrapper[4678]: I1206 10:49:12.160518 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-n2j4m" podStartSLOduration=3.499014894 podStartE2EDuration="7.16046971s" podCreationTimestamp="2025-12-06 10:49:05 +0000 UTC" firstStartedPulling="2025-12-06 10:49:05.509239084 +0000 UTC m=+750.352670523" lastFinishedPulling="2025-12-06 10:49:09.17069388 +0000 UTC m=+754.014125339" observedRunningTime="2025-12-06 10:49:10.19293698 +0000 UTC m=+755.036368419" watchObservedRunningTime="2025-12-06 10:49:12.16046971 +0000 UTC m=+757.003901149" Dec 06 10:49:12 crc kubenswrapper[4678]: I1206 10:49:12.161765 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-7nc8k" podStartSLOduration=2.452180361 podStartE2EDuration="8.16175772s" podCreationTimestamp="2025-12-06 10:49:04 +0000 UTC" firstStartedPulling="2025-12-06 10:49:05.904123869 +0000 UTC m=+750.747555328" lastFinishedPulling="2025-12-06 10:49:11.613701248 +0000 UTC m=+756.457132687" observedRunningTime="2025-12-06 10:49:12.158005573 +0000 UTC m=+757.001437082" watchObservedRunningTime="2025-12-06 10:49:12.16175772 +0000 UTC m=+757.005189159" Dec 06 10:49:15 crc kubenswrapper[4678]: I1206 10:49:15.489474 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-n2j4m" Dec 06 10:49:15 crc kubenswrapper[4678]: I1206 10:49:15.787101 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:15 crc kubenswrapper[4678]: I1206 10:49:15.787161 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:15 crc kubenswrapper[4678]: I1206 10:49:15.793806 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:16 crc kubenswrapper[4678]: I1206 10:49:16.174799 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-c76489864-zvgk9" Dec 06 10:49:16 crc kubenswrapper[4678]: I1206 10:49:16.257994 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ll9jl"] Dec 06 10:49:25 crc kubenswrapper[4678]: I1206 10:49:25.994248 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-8gwjz" Dec 06 10:49:29 crc kubenswrapper[4678]: I1206 10:49:29.505764 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:49:29 crc kubenswrapper[4678]: I1206 10:49:29.506351 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:49:29 crc kubenswrapper[4678]: I1206 10:49:29.506411 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:49:29 crc kubenswrapper[4678]: I1206 10:49:29.507145 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74cc7443fbc9fd2e3762e8b7e18efbd09edd85649bb12e6ad42344d0b88b6e39"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 10:49:29 crc kubenswrapper[4678]: I1206 10:49:29.507222 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://74cc7443fbc9fd2e3762e8b7e18efbd09edd85649bb12e6ad42344d0b88b6e39" gracePeriod=600 Dec 06 10:49:30 crc kubenswrapper[4678]: I1206 10:49:30.269420 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="74cc7443fbc9fd2e3762e8b7e18efbd09edd85649bb12e6ad42344d0b88b6e39" exitCode=0 Dec 06 10:49:30 crc kubenswrapper[4678]: I1206 10:49:30.269513 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"74cc7443fbc9fd2e3762e8b7e18efbd09edd85649bb12e6ad42344d0b88b6e39"} Dec 06 10:49:30 crc kubenswrapper[4678]: I1206 10:49:30.269573 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"a520bd68028a84131872db9f2785f29074c8e74fb3018d3a7e8874376510222f"} Dec 06 10:49:30 crc kubenswrapper[4678]: I1206 10:49:30.269596 4678 scope.go:117] "RemoveContainer" containerID="ae3c786b6ff131ce58d580d29415be4abad2449e4cbc0a482c5cc8f0f63c97cd" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.576457 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l"] Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.578616 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.581243 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.635236 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l"] Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.730007 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.730097 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ncgj\" (UniqueName: \"kubernetes.io/projected/ad0dea4c-c200-4a62-a5df-e718bffb082e-kube-api-access-8ncgj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.730371 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.832370 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.832443 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ncgj\" (UniqueName: \"kubernetes.io/projected/ad0dea4c-c200-4a62-a5df-e718bffb082e-kube-api-access-8ncgj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.832517 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.833111 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.833297 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.869975 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ncgj\" (UniqueName: \"kubernetes.io/projected/ad0dea4c-c200-4a62-a5df-e718bffb082e-kube-api-access-8ncgj\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:39 crc kubenswrapper[4678]: I1206 10:49:39.942977 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:40 crc kubenswrapper[4678]: I1206 10:49:40.195109 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l"] Dec 06 10:49:40 crc kubenswrapper[4678]: W1206 10:49:40.201043 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad0dea4c_c200_4a62_a5df_e718bffb082e.slice/crio-0d41f3eabbb998be50d82ee8919b11ee54feff30616f28c6277d896c4c21f445 WatchSource:0}: Error finding container 0d41f3eabbb998be50d82ee8919b11ee54feff30616f28c6277d896c4c21f445: Status 404 returned error can't find the container with id 0d41f3eabbb998be50d82ee8919b11ee54feff30616f28c6277d896c4c21f445 Dec 06 10:49:40 crc kubenswrapper[4678]: I1206 10:49:40.340760 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" event={"ID":"ad0dea4c-c200-4a62-a5df-e718bffb082e","Type":"ContainerStarted","Data":"0d41f3eabbb998be50d82ee8919b11ee54feff30616f28c6277d896c4c21f445"} Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.271827 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w24w9"] Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.273802 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.292317 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w24w9"] Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.331869 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ll9jl" podUID="f13f288c-c8dc-4df8-b89b-c325d088b30b" containerName="console" containerID="cri-o://fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960" gracePeriod=15 Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.352373 4678 generic.go:334] "Generic (PLEG): container finished" podID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerID="d49f633c51638ca186e135966ecaded3f80e17a6a7f60de0b51ff3949aa5e56b" exitCode=0 Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.352681 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" event={"ID":"ad0dea4c-c200-4a62-a5df-e718bffb082e","Type":"ContainerDied","Data":"d49f633c51638ca186e135966ecaded3f80e17a6a7f60de0b51ff3949aa5e56b"} Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.355315 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-catalog-content\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.355586 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdnqk\" (UniqueName: \"kubernetes.io/projected/7d6892a6-5947-4e64-b112-60b10ed2980c-kube-api-access-jdnqk\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.355725 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-utilities\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.457575 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-catalog-content\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.457750 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdnqk\" (UniqueName: \"kubernetes.io/projected/7d6892a6-5947-4e64-b112-60b10ed2980c-kube-api-access-jdnqk\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.457887 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-utilities\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.458620 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-utilities\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.459089 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-catalog-content\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.491330 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdnqk\" (UniqueName: \"kubernetes.io/projected/7d6892a6-5947-4e64-b112-60b10ed2980c-kube-api-access-jdnqk\") pod \"redhat-operators-w24w9\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.599862 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.796525 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ll9jl_f13f288c-c8dc-4df8-b89b-c325d088b30b/console/0.log" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.796604 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.895753 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w24w9"] Dec 06 10:49:41 crc kubenswrapper[4678]: W1206 10:49:41.903718 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d6892a6_5947_4e64_b112_60b10ed2980c.slice/crio-f6690954d4ea9d476d319ac0c50f7e4e05b75719cc7612b8de2a10b6e9d839e1 WatchSource:0}: Error finding container f6690954d4ea9d476d319ac0c50f7e4e05b75719cc7612b8de2a10b6e9d839e1: Status 404 returned error can't find the container with id f6690954d4ea9d476d319ac0c50f7e4e05b75719cc7612b8de2a10b6e9d839e1 Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.970869 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-trusted-ca-bundle\") pod \"f13f288c-c8dc-4df8-b89b-c325d088b30b\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.970926 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-oauth-config\") pod \"f13f288c-c8dc-4df8-b89b-c325d088b30b\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.970979 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvhww\" (UniqueName: \"kubernetes.io/projected/f13f288c-c8dc-4df8-b89b-c325d088b30b-kube-api-access-lvhww\") pod \"f13f288c-c8dc-4df8-b89b-c325d088b30b\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.971010 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-oauth-serving-cert\") pod \"f13f288c-c8dc-4df8-b89b-c325d088b30b\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.971030 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-config\") pod \"f13f288c-c8dc-4df8-b89b-c325d088b30b\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.971143 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-serving-cert\") pod \"f13f288c-c8dc-4df8-b89b-c325d088b30b\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.971168 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-service-ca\") pod \"f13f288c-c8dc-4df8-b89b-c325d088b30b\" (UID: \"f13f288c-c8dc-4df8-b89b-c325d088b30b\") " Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.971947 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f13f288c-c8dc-4df8-b89b-c325d088b30b" (UID: "f13f288c-c8dc-4df8-b89b-c325d088b30b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.971948 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f13f288c-c8dc-4df8-b89b-c325d088b30b" (UID: "f13f288c-c8dc-4df8-b89b-c325d088b30b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.972125 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-service-ca" (OuterVolumeSpecName: "service-ca") pod "f13f288c-c8dc-4df8-b89b-c325d088b30b" (UID: "f13f288c-c8dc-4df8-b89b-c325d088b30b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.972288 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-config" (OuterVolumeSpecName: "console-config") pod "f13f288c-c8dc-4df8-b89b-c325d088b30b" (UID: "f13f288c-c8dc-4df8-b89b-c325d088b30b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.979390 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f13f288c-c8dc-4df8-b89b-c325d088b30b" (UID: "f13f288c-c8dc-4df8-b89b-c325d088b30b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.979843 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f13f288c-c8dc-4df8-b89b-c325d088b30b-kube-api-access-lvhww" (OuterVolumeSpecName: "kube-api-access-lvhww") pod "f13f288c-c8dc-4df8-b89b-c325d088b30b" (UID: "f13f288c-c8dc-4df8-b89b-c325d088b30b"). InnerVolumeSpecName "kube-api-access-lvhww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:49:41 crc kubenswrapper[4678]: I1206 10:49:41.981615 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f13f288c-c8dc-4df8-b89b-c325d088b30b" (UID: "f13f288c-c8dc-4df8-b89b-c325d088b30b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.073364 4678 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.073397 4678 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.073409 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvhww\" (UniqueName: \"kubernetes.io/projected/f13f288c-c8dc-4df8-b89b-c325d088b30b-kube-api-access-lvhww\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.073420 4678 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.073428 4678 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.073437 4678 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f13f288c-c8dc-4df8-b89b-c325d088b30b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.073446 4678 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f13f288c-c8dc-4df8-b89b-c325d088b30b-service-ca\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.360734 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ll9jl_f13f288c-c8dc-4df8-b89b-c325d088b30b/console/0.log" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.360822 4678 generic.go:334] "Generic (PLEG): container finished" podID="f13f288c-c8dc-4df8-b89b-c325d088b30b" containerID="fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960" exitCode=2 Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.360874 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll9jl" event={"ID":"f13f288c-c8dc-4df8-b89b-c325d088b30b","Type":"ContainerDied","Data":"fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960"} Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.360899 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ll9jl" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.360931 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ll9jl" event={"ID":"f13f288c-c8dc-4df8-b89b-c325d088b30b","Type":"ContainerDied","Data":"532bb9d4a6fd1b7984067a6d937ac24c9b025e891f50325fbfb3997a8c59ee20"} Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.360959 4678 scope.go:117] "RemoveContainer" containerID="fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.367140 4678 generic.go:334] "Generic (PLEG): container finished" podID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerID="f8a593f74a71d983eff24ce3ef6b3e988d4ab4e442351854a6cd9d00c5e653e3" exitCode=0 Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.367207 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w24w9" event={"ID":"7d6892a6-5947-4e64-b112-60b10ed2980c","Type":"ContainerDied","Data":"f8a593f74a71d983eff24ce3ef6b3e988d4ab4e442351854a6cd9d00c5e653e3"} Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.367248 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w24w9" event={"ID":"7d6892a6-5947-4e64-b112-60b10ed2980c","Type":"ContainerStarted","Data":"f6690954d4ea9d476d319ac0c50f7e4e05b75719cc7612b8de2a10b6e9d839e1"} Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.394707 4678 scope.go:117] "RemoveContainer" containerID="fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960" Dec 06 10:49:42 crc kubenswrapper[4678]: E1206 10:49:42.397108 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960\": container with ID starting with fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960 not found: ID does not exist" containerID="fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.397180 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960"} err="failed to get container status \"fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960\": rpc error: code = NotFound desc = could not find container \"fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960\": container with ID starting with fccc0c5e41ca2c5cf0229bce946efd5f4a0accadec25a8d89d124eadc9d53960 not found: ID does not exist" Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.421273 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ll9jl"] Dec 06 10:49:42 crc kubenswrapper[4678]: I1206 10:49:42.436867 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ll9jl"] Dec 06 10:49:43 crc kubenswrapper[4678]: I1206 10:49:43.377959 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w24w9" event={"ID":"7d6892a6-5947-4e64-b112-60b10ed2980c","Type":"ContainerStarted","Data":"c76f02dc5fb64e20a6535e32886a97c7fe4f69c065488da83a38af0d2257a2f2"} Dec 06 10:49:43 crc kubenswrapper[4678]: I1206 10:49:43.384418 4678 generic.go:334] "Generic (PLEG): container finished" podID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerID="164ea9146b510cb705eec15e47604b478a76545338bd63938b71bcc9bcb79757" exitCode=0 Dec 06 10:49:43 crc kubenswrapper[4678]: I1206 10:49:43.384471 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" event={"ID":"ad0dea4c-c200-4a62-a5df-e718bffb082e","Type":"ContainerDied","Data":"164ea9146b510cb705eec15e47604b478a76545338bd63938b71bcc9bcb79757"} Dec 06 10:49:43 crc kubenswrapper[4678]: I1206 10:49:43.484524 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f13f288c-c8dc-4df8-b89b-c325d088b30b" path="/var/lib/kubelet/pods/f13f288c-c8dc-4df8-b89b-c325d088b30b/volumes" Dec 06 10:49:44 crc kubenswrapper[4678]: I1206 10:49:44.396324 4678 generic.go:334] "Generic (PLEG): container finished" podID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerID="c76f02dc5fb64e20a6535e32886a97c7fe4f69c065488da83a38af0d2257a2f2" exitCode=0 Dec 06 10:49:44 crc kubenswrapper[4678]: I1206 10:49:44.396419 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w24w9" event={"ID":"7d6892a6-5947-4e64-b112-60b10ed2980c","Type":"ContainerDied","Data":"c76f02dc5fb64e20a6535e32886a97c7fe4f69c065488da83a38af0d2257a2f2"} Dec 06 10:49:44 crc kubenswrapper[4678]: I1206 10:49:44.403615 4678 generic.go:334] "Generic (PLEG): container finished" podID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerID="48e97f6dfed818ffe203ee8ac0b7951e1529c257c0711c6f014e08507bf0c9ac" exitCode=0 Dec 06 10:49:44 crc kubenswrapper[4678]: I1206 10:49:44.403665 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" event={"ID":"ad0dea4c-c200-4a62-a5df-e718bffb082e","Type":"ContainerDied","Data":"48e97f6dfed818ffe203ee8ac0b7951e1529c257c0711c6f014e08507bf0c9ac"} Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.416378 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w24w9" event={"ID":"7d6892a6-5947-4e64-b112-60b10ed2980c","Type":"ContainerStarted","Data":"78461b15137689fbd0d3a487f623d63e82d7a67c4227cdd2f9333050424aec56"} Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.697619 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.748686 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w24w9" podStartSLOduration=2.33525022 podStartE2EDuration="4.748661385s" podCreationTimestamp="2025-12-06 10:49:41 +0000 UTC" firstStartedPulling="2025-12-06 10:49:42.369709438 +0000 UTC m=+787.213140877" lastFinishedPulling="2025-12-06 10:49:44.783120593 +0000 UTC m=+789.626552042" observedRunningTime="2025-12-06 10:49:45.450267259 +0000 UTC m=+790.293698708" watchObservedRunningTime="2025-12-06 10:49:45.748661385 +0000 UTC m=+790.592092834" Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.837626 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-bundle\") pod \"ad0dea4c-c200-4a62-a5df-e718bffb082e\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.837716 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-util\") pod \"ad0dea4c-c200-4a62-a5df-e718bffb082e\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.837831 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ncgj\" (UniqueName: \"kubernetes.io/projected/ad0dea4c-c200-4a62-a5df-e718bffb082e-kube-api-access-8ncgj\") pod \"ad0dea4c-c200-4a62-a5df-e718bffb082e\" (UID: \"ad0dea4c-c200-4a62-a5df-e718bffb082e\") " Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.838681 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-bundle" (OuterVolumeSpecName: "bundle") pod "ad0dea4c-c200-4a62-a5df-e718bffb082e" (UID: "ad0dea4c-c200-4a62-a5df-e718bffb082e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.845437 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0dea4c-c200-4a62-a5df-e718bffb082e-kube-api-access-8ncgj" (OuterVolumeSpecName: "kube-api-access-8ncgj") pod "ad0dea4c-c200-4a62-a5df-e718bffb082e" (UID: "ad0dea4c-c200-4a62-a5df-e718bffb082e"). InnerVolumeSpecName "kube-api-access-8ncgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.852778 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-util" (OuterVolumeSpecName: "util") pod "ad0dea4c-c200-4a62-a5df-e718bffb082e" (UID: "ad0dea4c-c200-4a62-a5df-e718bffb082e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.939578 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ncgj\" (UniqueName: \"kubernetes.io/projected/ad0dea4c-c200-4a62-a5df-e718bffb082e-kube-api-access-8ncgj\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.939625 4678 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:45 crc kubenswrapper[4678]: I1206 10:49:45.939638 4678 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/ad0dea4c-c200-4a62-a5df-e718bffb082e-util\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:46 crc kubenswrapper[4678]: I1206 10:49:46.431076 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" event={"ID":"ad0dea4c-c200-4a62-a5df-e718bffb082e","Type":"ContainerDied","Data":"0d41f3eabbb998be50d82ee8919b11ee54feff30616f28c6277d896c4c21f445"} Dec 06 10:49:46 crc kubenswrapper[4678]: I1206 10:49:46.431167 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d41f3eabbb998be50d82ee8919b11ee54feff30616f28c6277d896c4c21f445" Dec 06 10:49:46 crc kubenswrapper[4678]: I1206 10:49:46.431130 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l" Dec 06 10:49:51 crc kubenswrapper[4678]: I1206 10:49:51.601157 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:51 crc kubenswrapper[4678]: I1206 10:49:51.602034 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:51 crc kubenswrapper[4678]: I1206 10:49:51.649591 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:52 crc kubenswrapper[4678]: I1206 10:49:52.533374 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:53 crc kubenswrapper[4678]: I1206 10:49:53.653638 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w24w9"] Dec 06 10:49:54 crc kubenswrapper[4678]: I1206 10:49:54.479923 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w24w9" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="registry-server" containerID="cri-o://78461b15137689fbd0d3a487f623d63e82d7a67c4227cdd2f9333050424aec56" gracePeriod=2 Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.544750 4678 generic.go:334] "Generic (PLEG): container finished" podID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerID="78461b15137689fbd0d3a487f623d63e82d7a67c4227cdd2f9333050424aec56" exitCode=0 Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.545229 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w24w9" event={"ID":"7d6892a6-5947-4e64-b112-60b10ed2980c","Type":"ContainerDied","Data":"78461b15137689fbd0d3a487f623d63e82d7a67c4227cdd2f9333050424aec56"} Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.741805 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.782975 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-catalog-content\") pod \"7d6892a6-5947-4e64-b112-60b10ed2980c\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.783249 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-utilities\") pod \"7d6892a6-5947-4e64-b112-60b10ed2980c\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.783338 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdnqk\" (UniqueName: \"kubernetes.io/projected/7d6892a6-5947-4e64-b112-60b10ed2980c-kube-api-access-jdnqk\") pod \"7d6892a6-5947-4e64-b112-60b10ed2980c\" (UID: \"7d6892a6-5947-4e64-b112-60b10ed2980c\") " Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.785600 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-utilities" (OuterVolumeSpecName: "utilities") pod "7d6892a6-5947-4e64-b112-60b10ed2980c" (UID: "7d6892a6-5947-4e64-b112-60b10ed2980c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.789591 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d6892a6-5947-4e64-b112-60b10ed2980c-kube-api-access-jdnqk" (OuterVolumeSpecName: "kube-api-access-jdnqk") pod "7d6892a6-5947-4e64-b112-60b10ed2980c" (UID: "7d6892a6-5947-4e64-b112-60b10ed2980c"). InnerVolumeSpecName "kube-api-access-jdnqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.885553 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdnqk\" (UniqueName: \"kubernetes.io/projected/7d6892a6-5947-4e64-b112-60b10ed2980c-kube-api-access-jdnqk\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.885856 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.927946 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d6892a6-5947-4e64-b112-60b10ed2980c" (UID: "7d6892a6-5947-4e64-b112-60b10ed2980c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:49:56 crc kubenswrapper[4678]: I1206 10:49:56.986656 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d6892a6-5947-4e64-b112-60b10ed2980c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.212077 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-dd988988c-fc549"] Dec 06 10:49:57 crc kubenswrapper[4678]: E1206 10:49:57.212590 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="registry-server" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.212671 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="registry-server" Dec 06 10:49:57 crc kubenswrapper[4678]: E1206 10:49:57.212733 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f13f288c-c8dc-4df8-b89b-c325d088b30b" containerName="console" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.212783 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f13f288c-c8dc-4df8-b89b-c325d088b30b" containerName="console" Dec 06 10:49:57 crc kubenswrapper[4678]: E1206 10:49:57.212872 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerName="util" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.212930 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerName="util" Dec 06 10:49:57 crc kubenswrapper[4678]: E1206 10:49:57.213010 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerName="pull" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.213066 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerName="pull" Dec 06 10:49:57 crc kubenswrapper[4678]: E1206 10:49:57.213120 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="extract-content" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.213175 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="extract-content" Dec 06 10:49:57 crc kubenswrapper[4678]: E1206 10:49:57.213238 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerName="extract" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.213286 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerName="extract" Dec 06 10:49:57 crc kubenswrapper[4678]: E1206 10:49:57.213350 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="extract-utilities" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.213402 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="extract-utilities" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.213574 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0dea4c-c200-4a62-a5df-e718bffb082e" containerName="extract" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.213639 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f13f288c-c8dc-4df8-b89b-c325d088b30b" containerName="console" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.213711 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" containerName="registry-server" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.214199 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.219112 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.219210 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.226013 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.226184 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-nbpqz" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.229467 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.249223 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-dd988988c-fc549"] Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.289967 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc881548-41a9-4790-b691-dfb5935d128f-webhook-cert\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.290089 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc881548-41a9-4790-b691-dfb5935d128f-apiservice-cert\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.290126 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6hlm\" (UniqueName: \"kubernetes.io/projected/bc881548-41a9-4790-b691-dfb5935d128f-kube-api-access-g6hlm\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.391136 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc881548-41a9-4790-b691-dfb5935d128f-apiservice-cert\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.391189 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6hlm\" (UniqueName: \"kubernetes.io/projected/bc881548-41a9-4790-b691-dfb5935d128f-kube-api-access-g6hlm\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.391247 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc881548-41a9-4790-b691-dfb5935d128f-webhook-cert\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.397705 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bc881548-41a9-4790-b691-dfb5935d128f-webhook-cert\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.397702 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bc881548-41a9-4790-b691-dfb5935d128f-apiservice-cert\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.410210 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6hlm\" (UniqueName: \"kubernetes.io/projected/bc881548-41a9-4790-b691-dfb5935d128f-kube-api-access-g6hlm\") pod \"metallb-operator-controller-manager-dd988988c-fc549\" (UID: \"bc881548-41a9-4790-b691-dfb5935d128f\") " pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.529778 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.558389 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w24w9" event={"ID":"7d6892a6-5947-4e64-b112-60b10ed2980c","Type":"ContainerDied","Data":"f6690954d4ea9d476d319ac0c50f7e4e05b75719cc7612b8de2a10b6e9d839e1"} Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.558467 4678 scope.go:117] "RemoveContainer" containerID="78461b15137689fbd0d3a487f623d63e82d7a67c4227cdd2f9333050424aec56" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.558512 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w24w9" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.615643 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w24w9"] Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.657880 4678 scope.go:117] "RemoveContainer" containerID="c76f02dc5fb64e20a6535e32886a97c7fe4f69c065488da83a38af0d2257a2f2" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.658970 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w24w9"] Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.666617 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-d6f755774-dghr8"] Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.667410 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.669266 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d6f755774-dghr8"] Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.674251 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.674624 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.674795 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-v4242" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.716929 4678 scope.go:117] "RemoveContainer" containerID="f8a593f74a71d983eff24ce3ef6b3e988d4ab4e442351854a6cd9d00c5e653e3" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.812651 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8kmf\" (UniqueName: \"kubernetes.io/projected/2562672a-fcdf-420a-a4df-116113c240a0-kube-api-access-p8kmf\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.812985 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2562672a-fcdf-420a-a4df-116113c240a0-webhook-cert\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.813093 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2562672a-fcdf-420a-a4df-116113c240a0-apiservice-cert\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.914544 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2562672a-fcdf-420a-a4df-116113c240a0-apiservice-cert\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.914626 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8kmf\" (UniqueName: \"kubernetes.io/projected/2562672a-fcdf-420a-a4df-116113c240a0-kube-api-access-p8kmf\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.914664 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2562672a-fcdf-420a-a4df-116113c240a0-webhook-cert\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.920138 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2562672a-fcdf-420a-a4df-116113c240a0-webhook-cert\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.925320 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2562672a-fcdf-420a-a4df-116113c240a0-apiservice-cert\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:57 crc kubenswrapper[4678]: I1206 10:49:57.944880 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8kmf\" (UniqueName: \"kubernetes.io/projected/2562672a-fcdf-420a-a4df-116113c240a0-kube-api-access-p8kmf\") pod \"metallb-operator-webhook-server-d6f755774-dghr8\" (UID: \"2562672a-fcdf-420a-a4df-116113c240a0\") " pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:58 crc kubenswrapper[4678]: I1206 10:49:58.072854 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:49:58 crc kubenswrapper[4678]: I1206 10:49:58.247546 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-dd988988c-fc549"] Dec 06 10:49:58 crc kubenswrapper[4678]: W1206 10:49:58.262083 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc881548_41a9_4790_b691_dfb5935d128f.slice/crio-15851fec388de29bbbe79f7d6f6b5425db7905e87f6a88b8fa4d398a777e4a4b WatchSource:0}: Error finding container 15851fec388de29bbbe79f7d6f6b5425db7905e87f6a88b8fa4d398a777e4a4b: Status 404 returned error can't find the container with id 15851fec388de29bbbe79f7d6f6b5425db7905e87f6a88b8fa4d398a777e4a4b Dec 06 10:49:58 crc kubenswrapper[4678]: I1206 10:49:58.549035 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d6f755774-dghr8"] Dec 06 10:49:58 crc kubenswrapper[4678]: W1206 10:49:58.551972 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2562672a_fcdf_420a_a4df_116113c240a0.slice/crio-9480413af95273f331a21633e68ebbd94c5347bdf627f2255aa37b1451abd6d8 WatchSource:0}: Error finding container 9480413af95273f331a21633e68ebbd94c5347bdf627f2255aa37b1451abd6d8: Status 404 returned error can't find the container with id 9480413af95273f331a21633e68ebbd94c5347bdf627f2255aa37b1451abd6d8 Dec 06 10:49:58 crc kubenswrapper[4678]: I1206 10:49:58.568188 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" event={"ID":"bc881548-41a9-4790-b691-dfb5935d128f","Type":"ContainerStarted","Data":"15851fec388de29bbbe79f7d6f6b5425db7905e87f6a88b8fa4d398a777e4a4b"} Dec 06 10:49:58 crc kubenswrapper[4678]: I1206 10:49:58.569265 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" event={"ID":"2562672a-fcdf-420a-a4df-116113c240a0","Type":"ContainerStarted","Data":"9480413af95273f331a21633e68ebbd94c5347bdf627f2255aa37b1451abd6d8"} Dec 06 10:49:59 crc kubenswrapper[4678]: I1206 10:49:59.489704 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d6892a6-5947-4e64-b112-60b10ed2980c" path="/var/lib/kubelet/pods/7d6892a6-5947-4e64-b112-60b10ed2980c/volumes" Dec 06 10:50:05 crc kubenswrapper[4678]: I1206 10:50:05.626276 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" event={"ID":"bc881548-41a9-4790-b691-dfb5935d128f","Type":"ContainerStarted","Data":"751ce9a540a0711a76d5c6867c344330987aa019d0fa8517e5cbdfc17c541ebe"} Dec 06 10:50:05 crc kubenswrapper[4678]: I1206 10:50:05.626886 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:50:05 crc kubenswrapper[4678]: I1206 10:50:05.628596 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" event={"ID":"2562672a-fcdf-420a-a4df-116113c240a0","Type":"ContainerStarted","Data":"634317add89622c30d18ae46f553e6e2cdd05cb87b21fd16449e0a43f888a21b"} Dec 06 10:50:05 crc kubenswrapper[4678]: I1206 10:50:05.628788 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:50:05 crc kubenswrapper[4678]: I1206 10:50:05.662591 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" podStartSLOduration=2.358153161 podStartE2EDuration="8.662572441s" podCreationTimestamp="2025-12-06 10:49:57 +0000 UTC" firstStartedPulling="2025-12-06 10:49:58.264810234 +0000 UTC m=+803.108241673" lastFinishedPulling="2025-12-06 10:50:04.569229514 +0000 UTC m=+809.412660953" observedRunningTime="2025-12-06 10:50:05.653935801 +0000 UTC m=+810.497367240" watchObservedRunningTime="2025-12-06 10:50:05.662572441 +0000 UTC m=+810.506003870" Dec 06 10:50:05 crc kubenswrapper[4678]: I1206 10:50:05.685474 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" podStartSLOduration=2.650423184 podStartE2EDuration="8.685453064s" podCreationTimestamp="2025-12-06 10:49:57 +0000 UTC" firstStartedPulling="2025-12-06 10:49:58.554764116 +0000 UTC m=+803.398195545" lastFinishedPulling="2025-12-06 10:50:04.589793986 +0000 UTC m=+809.433225425" observedRunningTime="2025-12-06 10:50:05.682001796 +0000 UTC m=+810.525433235" watchObservedRunningTime="2025-12-06 10:50:05.685453064 +0000 UTC m=+810.528884503" Dec 06 10:50:18 crc kubenswrapper[4678]: I1206 10:50:18.078607 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-d6f755774-dghr8" Dec 06 10:50:37 crc kubenswrapper[4678]: I1206 10:50:37.532828 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-dd988988c-fc549" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.450075 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-n2dhr"] Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.453670 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.456480 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.458171 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.458462 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-hdmsv" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.463442 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m"] Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.464276 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.474353 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.499863 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b5c6\" (UniqueName: \"kubernetes.io/projected/1582c836-cc2c-4d01-b878-6fd5cde0ae70-kube-api-access-8b5c6\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.499930 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.508983 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m"] Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601469 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-conf\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601548 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-startup\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601609 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics-certs\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b5c6\" (UniqueName: \"kubernetes.io/projected/1582c836-cc2c-4d01-b878-6fd5cde0ae70-kube-api-access-8b5c6\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601760 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601781 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-reloader\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601806 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-sockets\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.601848 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmtkr\" (UniqueName: \"kubernetes.io/projected/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-kube-api-access-jmtkr\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.602385 4678 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.602447 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert podName:1582c836-cc2c-4d01-b878-6fd5cde0ae70 nodeName:}" failed. No retries permitted until 2025-12-06 10:50:39.102427083 +0000 UTC m=+843.945858522 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert") pod "frr-k8s-webhook-server-7fcb986d4-sqk7m" (UID: "1582c836-cc2c-4d01-b878-6fd5cde0ae70") : secret "frr-k8s-webhook-server-cert" not found Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.649796 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b5c6\" (UniqueName: \"kubernetes.io/projected/1582c836-cc2c-4d01-b878-6fd5cde0ae70-kube-api-access-8b5c6\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.670908 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qjlkq"] Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.672313 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.680066 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5kjrm" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.680326 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.680433 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.680598 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.705079 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57tvh\" (UniqueName: \"kubernetes.io/projected/e6542baf-28b2-4844-991e-eb0f57d45231-kube-api-access-57tvh\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.705370 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.705448 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e6542baf-28b2-4844-991e-eb0f57d45231-metallb-excludel2\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.705571 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.705664 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-reloader\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.706548 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-sockets\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.718055 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmtkr\" (UniqueName: \"kubernetes.io/projected/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-kube-api-access-jmtkr\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.718150 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-conf\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.718193 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-startup\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.718211 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics-certs\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.718266 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-metrics-certs\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.706482 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.712112 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-wstfq"] Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.719721 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.720599 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-conf\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.706473 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-reloader\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.721451 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-startup\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.712519 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-frr-sockets\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.723788 4678 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.727426 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics-certs podName:257b8a98-1ec9-4e1e-aef6-4f97cbe50546 nodeName:}" failed. No retries permitted until 2025-12-06 10:50:39.227383177 +0000 UTC m=+844.070814616 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics-certs") pod "frr-k8s-n2dhr" (UID: "257b8a98-1ec9-4e1e-aef6-4f97cbe50546") : secret "frr-k8s-certs-secret" not found Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.730138 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.754157 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-wstfq"] Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.789715 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmtkr\" (UniqueName: \"kubernetes.io/projected/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-kube-api-access-jmtkr\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.820212 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-metrics-certs\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.820274 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7cjs\" (UniqueName: \"kubernetes.io/projected/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-kube-api-access-p7cjs\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.820312 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.820338 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57tvh\" (UniqueName: \"kubernetes.io/projected/e6542baf-28b2-4844-991e-eb0f57d45231-kube-api-access-57tvh\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.820356 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e6542baf-28b2-4844-991e-eb0f57d45231-metallb-excludel2\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.820381 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-metrics-certs\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.820398 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-cert\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.820580 4678 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.820644 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-metrics-certs podName:e6542baf-28b2-4844-991e-eb0f57d45231 nodeName:}" failed. No retries permitted until 2025-12-06 10:50:39.320625391 +0000 UTC m=+844.164056830 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-metrics-certs") pod "speaker-qjlkq" (UID: "e6542baf-28b2-4844-991e-eb0f57d45231") : secret "speaker-certs-secret" not found Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.820890 4678 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.820926 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist podName:e6542baf-28b2-4844-991e-eb0f57d45231 nodeName:}" failed. No retries permitted until 2025-12-06 10:50:39.32091645 +0000 UTC m=+844.164347889 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist") pod "speaker-qjlkq" (UID: "e6542baf-28b2-4844-991e-eb0f57d45231") : secret "metallb-memberlist" not found Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.821728 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e6542baf-28b2-4844-991e-eb0f57d45231-metallb-excludel2\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.860605 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57tvh\" (UniqueName: \"kubernetes.io/projected/e6542baf-28b2-4844-991e-eb0f57d45231-kube-api-access-57tvh\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.921853 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-cert\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.921978 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7cjs\" (UniqueName: \"kubernetes.io/projected/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-kube-api-access-p7cjs\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.922028 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-metrics-certs\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.922152 4678 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 06 10:50:38 crc kubenswrapper[4678]: E1206 10:50:38.922210 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-metrics-certs podName:714f1cf8-6f0e-4c0f-8924-4b07a98c578f nodeName:}" failed. No retries permitted until 2025-12-06 10:50:39.422192508 +0000 UTC m=+844.265623937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-metrics-certs") pod "controller-f8648f98b-wstfq" (UID: "714f1cf8-6f0e-4c0f-8924-4b07a98c578f") : secret "controller-certs-secret" not found Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.928728 4678 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.936532 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-cert\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:38 crc kubenswrapper[4678]: I1206 10:50:38.965155 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7cjs\" (UniqueName: \"kubernetes.io/projected/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-kube-api-access-p7cjs\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.125105 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:39 crc kubenswrapper[4678]: E1206 10:50:39.125304 4678 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 06 10:50:39 crc kubenswrapper[4678]: E1206 10:50:39.125676 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert podName:1582c836-cc2c-4d01-b878-6fd5cde0ae70 nodeName:}" failed. No retries permitted until 2025-12-06 10:50:40.125653801 +0000 UTC m=+844.969085240 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert") pod "frr-k8s-webhook-server-7fcb986d4-sqk7m" (UID: "1582c836-cc2c-4d01-b878-6fd5cde0ae70") : secret "frr-k8s-webhook-server-cert" not found Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.328234 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-metrics-certs\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.328311 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.328394 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics-certs\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:39 crc kubenswrapper[4678]: E1206 10:50:39.328544 4678 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 10:50:39 crc kubenswrapper[4678]: E1206 10:50:39.328628 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist podName:e6542baf-28b2-4844-991e-eb0f57d45231 nodeName:}" failed. No retries permitted until 2025-12-06 10:50:40.32860272 +0000 UTC m=+845.172034159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist") pod "speaker-qjlkq" (UID: "e6542baf-28b2-4844-991e-eb0f57d45231") : secret "metallb-memberlist" not found Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.332141 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/257b8a98-1ec9-4e1e-aef6-4f97cbe50546-metrics-certs\") pod \"frr-k8s-n2dhr\" (UID: \"257b8a98-1ec9-4e1e-aef6-4f97cbe50546\") " pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.332558 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-metrics-certs\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.377759 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.429594 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-metrics-certs\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.433087 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/714f1cf8-6f0e-4c0f-8924-4b07a98c578f-metrics-certs\") pod \"controller-f8648f98b-wstfq\" (UID: \"714f1cf8-6f0e-4c0f-8924-4b07a98c578f\") " pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.675080 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:39 crc kubenswrapper[4678]: I1206 10:50:39.890852 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerStarted","Data":"bcccec7b2842ac59c82d590d49dc08cd0ab20b871c0642f4a272fc73c8cbed94"} Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.059134 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-wstfq"] Dec 06 10:50:40 crc kubenswrapper[4678]: W1206 10:50:40.064768 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod714f1cf8_6f0e_4c0f_8924_4b07a98c578f.slice/crio-67e7c172a5e4596b71446219dd6746581d1987b7d53eea2e195251eb8a2449a2 WatchSource:0}: Error finding container 67e7c172a5e4596b71446219dd6746581d1987b7d53eea2e195251eb8a2449a2: Status 404 returned error can't find the container with id 67e7c172a5e4596b71446219dd6746581d1987b7d53eea2e195251eb8a2449a2 Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.142723 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.151361 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1582c836-cc2c-4d01-b878-6fd5cde0ae70-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-sqk7m\" (UID: \"1582c836-cc2c-4d01-b878-6fd5cde0ae70\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.284263 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.346707 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:40 crc kubenswrapper[4678]: E1206 10:50:40.346956 4678 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 06 10:50:40 crc kubenswrapper[4678]: E1206 10:50:40.347023 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist podName:e6542baf-28b2-4844-991e-eb0f57d45231 nodeName:}" failed. No retries permitted until 2025-12-06 10:50:42.347007913 +0000 UTC m=+847.190439352 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist") pod "speaker-qjlkq" (UID: "e6542baf-28b2-4844-991e-eb0f57d45231") : secret "metallb-memberlist" not found Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.913469 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m"] Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.915665 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wstfq" event={"ID":"714f1cf8-6f0e-4c0f-8924-4b07a98c578f","Type":"ContainerStarted","Data":"7803bbe36253e6a7790dda82d77f7a1094cc0a5ee60fe9b8f70f9f8fe3ae34fd"} Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.915741 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wstfq" event={"ID":"714f1cf8-6f0e-4c0f-8924-4b07a98c578f","Type":"ContainerStarted","Data":"0df0712c12b1cf71f5d7994ce05d51091cabf5da5fbadfb2907b53b67cd4e95a"} Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.915759 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-wstfq" event={"ID":"714f1cf8-6f0e-4c0f-8924-4b07a98c578f","Type":"ContainerStarted","Data":"67e7c172a5e4596b71446219dd6746581d1987b7d53eea2e195251eb8a2449a2"} Dec 06 10:50:40 crc kubenswrapper[4678]: I1206 10:50:40.915803 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:40 crc kubenswrapper[4678]: W1206 10:50:40.931303 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1582c836_cc2c_4d01_b878_6fd5cde0ae70.slice/crio-7ce78928f98385f9ddf56642e20b644e08075eaf3bec26fbed1b81c744a17f4b WatchSource:0}: Error finding container 7ce78928f98385f9ddf56642e20b644e08075eaf3bec26fbed1b81c744a17f4b: Status 404 returned error can't find the container with id 7ce78928f98385f9ddf56642e20b644e08075eaf3bec26fbed1b81c744a17f4b Dec 06 10:50:41 crc kubenswrapper[4678]: I1206 10:50:41.000708 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-wstfq" podStartSLOduration=3.000688728 podStartE2EDuration="3.000688728s" podCreationTimestamp="2025-12-06 10:50:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:50:41.000128971 +0000 UTC m=+845.843560420" watchObservedRunningTime="2025-12-06 10:50:41.000688728 +0000 UTC m=+845.844120167" Dec 06 10:50:41 crc kubenswrapper[4678]: I1206 10:50:41.927038 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" event={"ID":"1582c836-cc2c-4d01-b878-6fd5cde0ae70","Type":"ContainerStarted","Data":"7ce78928f98385f9ddf56642e20b644e08075eaf3bec26fbed1b81c744a17f4b"} Dec 06 10:50:42 crc kubenswrapper[4678]: I1206 10:50:42.407427 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:42 crc kubenswrapper[4678]: I1206 10:50:42.431875 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e6542baf-28b2-4844-991e-eb0f57d45231-memberlist\") pod \"speaker-qjlkq\" (UID: \"e6542baf-28b2-4844-991e-eb0f57d45231\") " pod="metallb-system/speaker-qjlkq" Dec 06 10:50:42 crc kubenswrapper[4678]: I1206 10:50:42.604878 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qjlkq" Dec 06 10:50:42 crc kubenswrapper[4678]: I1206 10:50:42.954540 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qjlkq" event={"ID":"e6542baf-28b2-4844-991e-eb0f57d45231","Type":"ContainerStarted","Data":"105c6be702cea5e4585e2fdcaf4120f142f907516b8e215f1d8ac089b6bc33b7"} Dec 06 10:50:43 crc kubenswrapper[4678]: I1206 10:50:43.968160 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qjlkq" event={"ID":"e6542baf-28b2-4844-991e-eb0f57d45231","Type":"ContainerStarted","Data":"0c2b6306ba7b66d1e5203219554b623dced9eb3ac40232d122744ac400fd9719"} Dec 06 10:50:43 crc kubenswrapper[4678]: I1206 10:50:43.968211 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qjlkq" event={"ID":"e6542baf-28b2-4844-991e-eb0f57d45231","Type":"ContainerStarted","Data":"e0f61b747289c3b0f4ce4857b52c2230187952165227773b1bda6560c2ed3dcd"} Dec 06 10:50:43 crc kubenswrapper[4678]: I1206 10:50:43.968298 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qjlkq" Dec 06 10:50:44 crc kubenswrapper[4678]: I1206 10:50:44.000511 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qjlkq" podStartSLOduration=6.000474835 podStartE2EDuration="6.000474835s" podCreationTimestamp="2025-12-06 10:50:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:50:43.995810399 +0000 UTC m=+848.839241838" watchObservedRunningTime="2025-12-06 10:50:44.000474835 +0000 UTC m=+848.843906274" Dec 06 10:50:50 crc kubenswrapper[4678]: I1206 10:50:50.028645 4678 generic.go:334] "Generic (PLEG): container finished" podID="257b8a98-1ec9-4e1e-aef6-4f97cbe50546" containerID="c6244c9516486f29dd32399f98b6eb60d2c43391ba74b8e7d5d2c4805e327bdc" exitCode=0 Dec 06 10:50:50 crc kubenswrapper[4678]: I1206 10:50:50.028739 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerDied","Data":"c6244c9516486f29dd32399f98b6eb60d2c43391ba74b8e7d5d2c4805e327bdc"} Dec 06 10:50:50 crc kubenswrapper[4678]: I1206 10:50:50.030774 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" event={"ID":"1582c836-cc2c-4d01-b878-6fd5cde0ae70","Type":"ContainerStarted","Data":"6877e26597baa74b8d56a051c8131b50440aeb5a5864ba341a3a5d9d1b12bbc2"} Dec 06 10:50:50 crc kubenswrapper[4678]: I1206 10:50:50.030951 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:50:50 crc kubenswrapper[4678]: I1206 10:50:50.053809 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" podStartSLOduration=3.540626522 podStartE2EDuration="12.053783621s" podCreationTimestamp="2025-12-06 10:50:38 +0000 UTC" firstStartedPulling="2025-12-06 10:50:40.934431702 +0000 UTC m=+845.777863141" lastFinishedPulling="2025-12-06 10:50:49.447588811 +0000 UTC m=+854.291020240" observedRunningTime="2025-12-06 10:50:50.050246549 +0000 UTC m=+854.893677988" watchObservedRunningTime="2025-12-06 10:50:50.053783621 +0000 UTC m=+854.897215060" Dec 06 10:50:52 crc kubenswrapper[4678]: I1206 10:50:52.049422 4678 generic.go:334] "Generic (PLEG): container finished" podID="257b8a98-1ec9-4e1e-aef6-4f97cbe50546" containerID="8957caf046eab2ebbc263c9be980b8b605eae38579bbd2b61469b2e699e92ad5" exitCode=0 Dec 06 10:50:52 crc kubenswrapper[4678]: I1206 10:50:52.049484 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerDied","Data":"8957caf046eab2ebbc263c9be980b8b605eae38579bbd2b61469b2e699e92ad5"} Dec 06 10:50:52 crc kubenswrapper[4678]: I1206 10:50:52.609553 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qjlkq" Dec 06 10:50:53 crc kubenswrapper[4678]: I1206 10:50:53.058775 4678 generic.go:334] "Generic (PLEG): container finished" podID="257b8a98-1ec9-4e1e-aef6-4f97cbe50546" containerID="e6117631f690476be2ad5ef1bad8f92e6e09756d9d6987cfd66d424360c20fad" exitCode=0 Dec 06 10:50:53 crc kubenswrapper[4678]: I1206 10:50:53.058828 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerDied","Data":"e6117631f690476be2ad5ef1bad8f92e6e09756d9d6987cfd66d424360c20fad"} Dec 06 10:50:54 crc kubenswrapper[4678]: I1206 10:50:54.071025 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerStarted","Data":"98a85cb30f3f2157d92a50e713643a87c78e26769bad4a92435834f440e55181"} Dec 06 10:50:54 crc kubenswrapper[4678]: I1206 10:50:54.071599 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerStarted","Data":"edc7063cf5867e7ed0af1deb0c27daa2e6d3186c08f86f4036f1ba4093b3b812"} Dec 06 10:50:54 crc kubenswrapper[4678]: I1206 10:50:54.071616 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerStarted","Data":"32e20f63096eb7315db376515ed6230963cae685f54c1199a0e7a591243cea46"} Dec 06 10:50:54 crc kubenswrapper[4678]: I1206 10:50:54.071630 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerStarted","Data":"ade7663447e9dba9cd6ca2cb4c66aa9425a76621e58f6c731a04ce6c7c715885"} Dec 06 10:50:54 crc kubenswrapper[4678]: I1206 10:50:54.071643 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerStarted","Data":"a8265fedcd358c00418e294c7cf5216e60c01c1227e2fbdb230907b2a85fbac3"} Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.089032 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n2dhr" event={"ID":"257b8a98-1ec9-4e1e-aef6-4f97cbe50546","Type":"ContainerStarted","Data":"c16ed95c3cbe7e245ab623952610331e3d78ba20c2a62e85ed9182d4f0eb822a"} Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.089837 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.125767 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-n2dhr" podStartSLOduration=7.404807206 podStartE2EDuration="17.125737719s" podCreationTimestamp="2025-12-06 10:50:38 +0000 UTC" firstStartedPulling="2025-12-06 10:50:39.700396782 +0000 UTC m=+844.543828221" lastFinishedPulling="2025-12-06 10:50:49.421327285 +0000 UTC m=+854.264758734" observedRunningTime="2025-12-06 10:50:55.116675103 +0000 UTC m=+859.960106602" watchObservedRunningTime="2025-12-06 10:50:55.125737719 +0000 UTC m=+859.969169188" Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.839352 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-rbppr"] Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.840331 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rbppr" Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.850328 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-qlwm2" Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.850336 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.850513 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.863732 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rbppr"] Dec 06 10:50:55 crc kubenswrapper[4678]: I1206 10:50:55.938434 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q5bh\" (UniqueName: \"kubernetes.io/projected/f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8-kube-api-access-8q5bh\") pod \"openstack-operator-index-rbppr\" (UID: \"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8\") " pod="openstack-operators/openstack-operator-index-rbppr" Dec 06 10:50:56 crc kubenswrapper[4678]: I1206 10:50:56.040247 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q5bh\" (UniqueName: \"kubernetes.io/projected/f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8-kube-api-access-8q5bh\") pod \"openstack-operator-index-rbppr\" (UID: \"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8\") " pod="openstack-operators/openstack-operator-index-rbppr" Dec 06 10:50:56 crc kubenswrapper[4678]: I1206 10:50:56.072757 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q5bh\" (UniqueName: \"kubernetes.io/projected/f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8-kube-api-access-8q5bh\") pod \"openstack-operator-index-rbppr\" (UID: \"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8\") " pod="openstack-operators/openstack-operator-index-rbppr" Dec 06 10:50:56 crc kubenswrapper[4678]: I1206 10:50:56.167163 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rbppr" Dec 06 10:50:56 crc kubenswrapper[4678]: I1206 10:50:56.453645 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-rbppr"] Dec 06 10:50:57 crc kubenswrapper[4678]: I1206 10:50:57.103540 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rbppr" event={"ID":"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8","Type":"ContainerStarted","Data":"5e3f2453cf304e9bf7351fc28d666726815c4a7f9ff6f07eeede81e242aeabb5"} Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.120998 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rbppr" event={"ID":"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8","Type":"ContainerStarted","Data":"bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4"} Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.145571 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-rbppr" podStartSLOduration=1.8355686150000001 podStartE2EDuration="4.14554313s" podCreationTimestamp="2025-12-06 10:50:55 +0000 UTC" firstStartedPulling="2025-12-06 10:50:56.462214944 +0000 UTC m=+861.305646383" lastFinishedPulling="2025-12-06 10:50:58.772189459 +0000 UTC m=+863.615620898" observedRunningTime="2025-12-06 10:50:59.139681015 +0000 UTC m=+863.983112484" watchObservedRunningTime="2025-12-06 10:50:59.14554313 +0000 UTC m=+863.988974609" Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.196159 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rbppr"] Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.378362 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.423383 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.681193 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-wstfq" Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.803926 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-m7422"] Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.805197 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.816382 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m7422"] Dec 06 10:50:59 crc kubenswrapper[4678]: I1206 10:50:59.909791 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z48n2\" (UniqueName: \"kubernetes.io/projected/dfb6011a-3444-4e31-8a84-d2c521f55e77-kube-api-access-z48n2\") pod \"openstack-operator-index-m7422\" (UID: \"dfb6011a-3444-4e31-8a84-d2c521f55e77\") " pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:00 crc kubenswrapper[4678]: I1206 10:51:00.011618 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z48n2\" (UniqueName: \"kubernetes.io/projected/dfb6011a-3444-4e31-8a84-d2c521f55e77-kube-api-access-z48n2\") pod \"openstack-operator-index-m7422\" (UID: \"dfb6011a-3444-4e31-8a84-d2c521f55e77\") " pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:00 crc kubenswrapper[4678]: I1206 10:51:00.033140 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z48n2\" (UniqueName: \"kubernetes.io/projected/dfb6011a-3444-4e31-8a84-d2c521f55e77-kube-api-access-z48n2\") pod \"openstack-operator-index-m7422\" (UID: \"dfb6011a-3444-4e31-8a84-d2c521f55e77\") " pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:00 crc kubenswrapper[4678]: I1206 10:51:00.180457 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:00 crc kubenswrapper[4678]: I1206 10:51:00.295547 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-sqk7m" Dec 06 10:51:00 crc kubenswrapper[4678]: I1206 10:51:00.519680 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m7422"] Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.144032 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m7422" event={"ID":"dfb6011a-3444-4e31-8a84-d2c521f55e77","Type":"ContainerStarted","Data":"a0649ef69e5368ec7f6882656bf7d5d23be1d5bdd36e142b429fb409b898dcc1"} Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.144900 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m7422" event={"ID":"dfb6011a-3444-4e31-8a84-d2c521f55e77","Type":"ContainerStarted","Data":"3b7ffe66bb4fc67a5a5ed834921e74589b9ee2be3b4d869886b7ee8406792fd0"} Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.144128 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-rbppr" podUID="f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8" containerName="registry-server" containerID="cri-o://bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4" gracePeriod=2 Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.163528 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-m7422" podStartSLOduration=2.095024571 podStartE2EDuration="2.163472925s" podCreationTimestamp="2025-12-06 10:50:59 +0000 UTC" firstStartedPulling="2025-12-06 10:51:00.56634213 +0000 UTC m=+865.409773569" lastFinishedPulling="2025-12-06 10:51:00.634790484 +0000 UTC m=+865.478221923" observedRunningTime="2025-12-06 10:51:01.163187155 +0000 UTC m=+866.006618594" watchObservedRunningTime="2025-12-06 10:51:01.163472925 +0000 UTC m=+866.006904354" Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.529963 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rbppr" Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.642175 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q5bh\" (UniqueName: \"kubernetes.io/projected/f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8-kube-api-access-8q5bh\") pod \"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8\" (UID: \"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8\") " Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.651726 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8-kube-api-access-8q5bh" (OuterVolumeSpecName: "kube-api-access-8q5bh") pod "f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8" (UID: "f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8"). InnerVolumeSpecName "kube-api-access-8q5bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:51:01 crc kubenswrapper[4678]: I1206 10:51:01.744233 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q5bh\" (UniqueName: \"kubernetes.io/projected/f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8-kube-api-access-8q5bh\") on node \"crc\" DevicePath \"\"" Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.152799 4678 generic.go:334] "Generic (PLEG): container finished" podID="f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8" containerID="bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4" exitCode=0 Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.153561 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-rbppr" Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.153571 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rbppr" event={"ID":"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8","Type":"ContainerDied","Data":"bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4"} Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.153685 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-rbppr" event={"ID":"f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8","Type":"ContainerDied","Data":"5e3f2453cf304e9bf7351fc28d666726815c4a7f9ff6f07eeede81e242aeabb5"} Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.153714 4678 scope.go:117] "RemoveContainer" containerID="bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4" Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.174809 4678 scope.go:117] "RemoveContainer" containerID="bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4" Dec 06 10:51:02 crc kubenswrapper[4678]: E1206 10:51:02.175395 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4\": container with ID starting with bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4 not found: ID does not exist" containerID="bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4" Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.175458 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4"} err="failed to get container status \"bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4\": rpc error: code = NotFound desc = could not find container \"bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4\": container with ID starting with bd4e152dbe078564d8ad4b7e56f112fe967a14fe44b1b77cbe578b447926e3c4 not found: ID does not exist" Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.188473 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-rbppr"] Dec 06 10:51:02 crc kubenswrapper[4678]: I1206 10:51:02.193721 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-rbppr"] Dec 06 10:51:03 crc kubenswrapper[4678]: I1206 10:51:03.484393 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8" path="/var/lib/kubelet/pods/f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8/volumes" Dec 06 10:51:09 crc kubenswrapper[4678]: I1206 10:51:09.385450 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-n2dhr" Dec 06 10:51:10 crc kubenswrapper[4678]: I1206 10:51:10.181188 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:10 crc kubenswrapper[4678]: I1206 10:51:10.181475 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:10 crc kubenswrapper[4678]: I1206 10:51:10.224736 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:11 crc kubenswrapper[4678]: I1206 10:51:11.258481 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-m7422" Dec 06 10:51:12 crc kubenswrapper[4678]: I1206 10:51:12.873923 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd"] Dec 06 10:51:12 crc kubenswrapper[4678]: E1206 10:51:12.874260 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8" containerName="registry-server" Dec 06 10:51:12 crc kubenswrapper[4678]: I1206 10:51:12.874278 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8" containerName="registry-server" Dec 06 10:51:12 crc kubenswrapper[4678]: I1206 10:51:12.874412 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f49f48ea-6bc8-40fe-b94d-b23ebe2ef0f8" containerName="registry-server" Dec 06 10:51:12 crc kubenswrapper[4678]: I1206 10:51:12.875407 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:12 crc kubenswrapper[4678]: I1206 10:51:12.879922 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wprr5" Dec 06 10:51:12 crc kubenswrapper[4678]: I1206 10:51:12.903298 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd"] Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.051063 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf9tp\" (UniqueName: \"kubernetes.io/projected/e517ee42-f060-4681-97fd-386010a0fc4d-kube-api-access-bf9tp\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.051182 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-bundle\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.051232 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-util\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.152735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf9tp\" (UniqueName: \"kubernetes.io/projected/e517ee42-f060-4681-97fd-386010a0fc4d-kube-api-access-bf9tp\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.153259 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-bundle\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.153406 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-util\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.153990 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-bundle\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.154239 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-util\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.188234 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf9tp\" (UniqueName: \"kubernetes.io/projected/e517ee42-f060-4681-97fd-386010a0fc4d-kube-api-access-bf9tp\") pod \"67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.200446 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:13 crc kubenswrapper[4678]: I1206 10:51:13.513107 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd"] Dec 06 10:51:13 crc kubenswrapper[4678]: W1206 10:51:13.515126 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode517ee42_f060_4681_97fd_386010a0fc4d.slice/crio-aaf8af7f4d3a3fe896e2a269f47de87d59829eb61351852177f8a61966527e9e WatchSource:0}: Error finding container aaf8af7f4d3a3fe896e2a269f47de87d59829eb61351852177f8a61966527e9e: Status 404 returned error can't find the container with id aaf8af7f4d3a3fe896e2a269f47de87d59829eb61351852177f8a61966527e9e Dec 06 10:51:14 crc kubenswrapper[4678]: I1206 10:51:14.256902 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" event={"ID":"e517ee42-f060-4681-97fd-386010a0fc4d","Type":"ContainerStarted","Data":"aaf8af7f4d3a3fe896e2a269f47de87d59829eb61351852177f8a61966527e9e"} Dec 06 10:51:15 crc kubenswrapper[4678]: I1206 10:51:15.268066 4678 generic.go:334] "Generic (PLEG): container finished" podID="e517ee42-f060-4681-97fd-386010a0fc4d" containerID="bde507cb13b002fb551cac53c517583c5578ace3d31b330f6e31103b962de483" exitCode=0 Dec 06 10:51:15 crc kubenswrapper[4678]: I1206 10:51:15.268152 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" event={"ID":"e517ee42-f060-4681-97fd-386010a0fc4d","Type":"ContainerDied","Data":"bde507cb13b002fb551cac53c517583c5578ace3d31b330f6e31103b962de483"} Dec 06 10:51:16 crc kubenswrapper[4678]: I1206 10:51:16.278025 4678 generic.go:334] "Generic (PLEG): container finished" podID="e517ee42-f060-4681-97fd-386010a0fc4d" containerID="87274210bf2437cd6e0571e6f917d1ea7399a0eecd97c433da8d81c766393daa" exitCode=0 Dec 06 10:51:16 crc kubenswrapper[4678]: I1206 10:51:16.278329 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" event={"ID":"e517ee42-f060-4681-97fd-386010a0fc4d","Type":"ContainerDied","Data":"87274210bf2437cd6e0571e6f917d1ea7399a0eecd97c433da8d81c766393daa"} Dec 06 10:51:17 crc kubenswrapper[4678]: I1206 10:51:17.290770 4678 generic.go:334] "Generic (PLEG): container finished" podID="e517ee42-f060-4681-97fd-386010a0fc4d" containerID="5150e0fad61fbfa6fa96792ad973afe92e874b762168121086594e371f5f3396" exitCode=0 Dec 06 10:51:17 crc kubenswrapper[4678]: I1206 10:51:17.290879 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" event={"ID":"e517ee42-f060-4681-97fd-386010a0fc4d","Type":"ContainerDied","Data":"5150e0fad61fbfa6fa96792ad973afe92e874b762168121086594e371f5f3396"} Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.628351 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.750801 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-util\") pod \"e517ee42-f060-4681-97fd-386010a0fc4d\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.750918 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf9tp\" (UniqueName: \"kubernetes.io/projected/e517ee42-f060-4681-97fd-386010a0fc4d-kube-api-access-bf9tp\") pod \"e517ee42-f060-4681-97fd-386010a0fc4d\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.750945 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-bundle\") pod \"e517ee42-f060-4681-97fd-386010a0fc4d\" (UID: \"e517ee42-f060-4681-97fd-386010a0fc4d\") " Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.751762 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-bundle" (OuterVolumeSpecName: "bundle") pod "e517ee42-f060-4681-97fd-386010a0fc4d" (UID: "e517ee42-f060-4681-97fd-386010a0fc4d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.759784 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e517ee42-f060-4681-97fd-386010a0fc4d-kube-api-access-bf9tp" (OuterVolumeSpecName: "kube-api-access-bf9tp") pod "e517ee42-f060-4681-97fd-386010a0fc4d" (UID: "e517ee42-f060-4681-97fd-386010a0fc4d"). InnerVolumeSpecName "kube-api-access-bf9tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.769145 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-util" (OuterVolumeSpecName: "util") pod "e517ee42-f060-4681-97fd-386010a0fc4d" (UID: "e517ee42-f060-4681-97fd-386010a0fc4d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.853146 4678 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-util\") on node \"crc\" DevicePath \"\"" Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.853211 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf9tp\" (UniqueName: \"kubernetes.io/projected/e517ee42-f060-4681-97fd-386010a0fc4d-kube-api-access-bf9tp\") on node \"crc\" DevicePath \"\"" Dec 06 10:51:18 crc kubenswrapper[4678]: I1206 10:51:18.853222 4678 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e517ee42-f060-4681-97fd-386010a0fc4d-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:51:19 crc kubenswrapper[4678]: I1206 10:51:19.312572 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" event={"ID":"e517ee42-f060-4681-97fd-386010a0fc4d","Type":"ContainerDied","Data":"aaf8af7f4d3a3fe896e2a269f47de87d59829eb61351852177f8a61966527e9e"} Dec 06 10:51:19 crc kubenswrapper[4678]: I1206 10:51:19.312664 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aaf8af7f4d3a3fe896e2a269f47de87d59829eb61351852177f8a61966527e9e" Dec 06 10:51:19 crc kubenswrapper[4678]: I1206 10:51:19.312763 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.624637 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs"] Dec 06 10:51:25 crc kubenswrapper[4678]: E1206 10:51:25.625295 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e517ee42-f060-4681-97fd-386010a0fc4d" containerName="pull" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.625310 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e517ee42-f060-4681-97fd-386010a0fc4d" containerName="pull" Dec 06 10:51:25 crc kubenswrapper[4678]: E1206 10:51:25.625319 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e517ee42-f060-4681-97fd-386010a0fc4d" containerName="extract" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.625326 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e517ee42-f060-4681-97fd-386010a0fc4d" containerName="extract" Dec 06 10:51:25 crc kubenswrapper[4678]: E1206 10:51:25.625337 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e517ee42-f060-4681-97fd-386010a0fc4d" containerName="util" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.625345 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e517ee42-f060-4681-97fd-386010a0fc4d" containerName="util" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.625474 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e517ee42-f060-4681-97fd-386010a0fc4d" containerName="extract" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.625994 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.628651 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-ll7hf" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.660141 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vkwj\" (UniqueName: \"kubernetes.io/projected/0b336e20-2347-4b86-86b4-22ef7da8983c-kube-api-access-6vkwj\") pod \"openstack-operator-controller-operator-5bbffdd548-5zgrs\" (UID: \"0b336e20-2347-4b86-86b4-22ef7da8983c\") " pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.673312 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs"] Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.761293 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vkwj\" (UniqueName: \"kubernetes.io/projected/0b336e20-2347-4b86-86b4-22ef7da8983c-kube-api-access-6vkwj\") pod \"openstack-operator-controller-operator-5bbffdd548-5zgrs\" (UID: \"0b336e20-2347-4b86-86b4-22ef7da8983c\") " pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.785420 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vkwj\" (UniqueName: \"kubernetes.io/projected/0b336e20-2347-4b86-86b4-22ef7da8983c-kube-api-access-6vkwj\") pod \"openstack-operator-controller-operator-5bbffdd548-5zgrs\" (UID: \"0b336e20-2347-4b86-86b4-22ef7da8983c\") " pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" Dec 06 10:51:25 crc kubenswrapper[4678]: I1206 10:51:25.949111 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" Dec 06 10:51:26 crc kubenswrapper[4678]: I1206 10:51:26.461724 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs"] Dec 06 10:51:27 crc kubenswrapper[4678]: I1206 10:51:27.395023 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" event={"ID":"0b336e20-2347-4b86-86b4-22ef7da8983c","Type":"ContainerStarted","Data":"666322ce78e13f44f3c8ed937f2e5a4af223acadd90c8b8a4f4eb05212535d0f"} Dec 06 10:51:29 crc kubenswrapper[4678]: I1206 10:51:29.505884 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:51:29 crc kubenswrapper[4678]: I1206 10:51:29.509652 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:51:31 crc kubenswrapper[4678]: I1206 10:51:31.428302 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" event={"ID":"0b336e20-2347-4b86-86b4-22ef7da8983c","Type":"ContainerStarted","Data":"8b3694318c64561a8400e51b1705e7aa6cbf42b63d3572d47e49bde876919218"} Dec 06 10:51:31 crc kubenswrapper[4678]: I1206 10:51:31.428835 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" Dec 06 10:51:31 crc kubenswrapper[4678]: I1206 10:51:31.466812 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" podStartSLOduration=2.157377302 podStartE2EDuration="6.46678962s" podCreationTimestamp="2025-12-06 10:51:25 +0000 UTC" firstStartedPulling="2025-12-06 10:51:26.489301765 +0000 UTC m=+891.332733204" lastFinishedPulling="2025-12-06 10:51:30.798714093 +0000 UTC m=+895.642145522" observedRunningTime="2025-12-06 10:51:31.460974037 +0000 UTC m=+896.304405496" watchObservedRunningTime="2025-12-06 10:51:31.46678962 +0000 UTC m=+896.310221069" Dec 06 10:51:35 crc kubenswrapper[4678]: I1206 10:51:35.952975 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5bbffdd548-5zgrs" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.097407 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nldgn"] Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.099509 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.110450 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-catalog-content\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.110886 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-utilities\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.110955 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzwqp\" (UniqueName: \"kubernetes.io/projected/f7e156c4-11f4-435c-bd15-431daf7c97c1-kube-api-access-qzwqp\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.123294 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nldgn"] Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.212589 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-utilities\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.212954 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzwqp\" (UniqueName: \"kubernetes.io/projected/f7e156c4-11f4-435c-bd15-431daf7c97c1-kube-api-access-qzwqp\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.213062 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-catalog-content\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.213576 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-catalog-content\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.213685 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-utilities\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.239902 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzwqp\" (UniqueName: \"kubernetes.io/projected/f7e156c4-11f4-435c-bd15-431daf7c97c1-kube-api-access-qzwqp\") pod \"certified-operators-nldgn\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.418841 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:47 crc kubenswrapper[4678]: I1206 10:51:47.828844 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nldgn"] Dec 06 10:51:48 crc kubenswrapper[4678]: I1206 10:51:48.541442 4678 generic.go:334] "Generic (PLEG): container finished" podID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerID="83243b8c01c294af1cd8ae3458e3d6c010e560cd9add00939742aa4303a793e0" exitCode=0 Dec 06 10:51:48 crc kubenswrapper[4678]: I1206 10:51:48.541862 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldgn" event={"ID":"f7e156c4-11f4-435c-bd15-431daf7c97c1","Type":"ContainerDied","Data":"83243b8c01c294af1cd8ae3458e3d6c010e560cd9add00939742aa4303a793e0"} Dec 06 10:51:48 crc kubenswrapper[4678]: I1206 10:51:48.541893 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldgn" event={"ID":"f7e156c4-11f4-435c-bd15-431daf7c97c1","Type":"ContainerStarted","Data":"dccbc2d254269a5773c1a28b817d72d99027458fd4d0028e90b49b18e35b8d82"} Dec 06 10:51:50 crc kubenswrapper[4678]: I1206 10:51:50.567378 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldgn" event={"ID":"f7e156c4-11f4-435c-bd15-431daf7c97c1","Type":"ContainerStarted","Data":"254bc4b474591f0e3849f86c4198a21cd58c274756c497618b470048e2028c07"} Dec 06 10:51:51 crc kubenswrapper[4678]: I1206 10:51:51.574378 4678 generic.go:334] "Generic (PLEG): container finished" podID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerID="254bc4b474591f0e3849f86c4198a21cd58c274756c497618b470048e2028c07" exitCode=0 Dec 06 10:51:51 crc kubenswrapper[4678]: I1206 10:51:51.574441 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldgn" event={"ID":"f7e156c4-11f4-435c-bd15-431daf7c97c1","Type":"ContainerDied","Data":"254bc4b474591f0e3849f86c4198a21cd58c274756c497618b470048e2028c07"} Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.584309 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldgn" event={"ID":"f7e156c4-11f4-435c-bd15-431daf7c97c1","Type":"ContainerStarted","Data":"76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4"} Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.606292 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nldgn" podStartSLOduration=1.913742906 podStartE2EDuration="5.606269317s" podCreationTimestamp="2025-12-06 10:51:47 +0000 UTC" firstStartedPulling="2025-12-06 10:51:48.543143733 +0000 UTC m=+913.386575172" lastFinishedPulling="2025-12-06 10:51:52.235670154 +0000 UTC m=+917.079101583" observedRunningTime="2025-12-06 10:51:52.605653529 +0000 UTC m=+917.449084968" watchObservedRunningTime="2025-12-06 10:51:52.606269317 +0000 UTC m=+917.449700786" Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.829319 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw"] Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.831002 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.838431 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hjgmn" Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.890781 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw"] Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.915750 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw"] Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.917023 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" Dec 06 10:51:52 crc kubenswrapper[4678]: W1206 10:51:52.919689 4678 reflector.go:561] object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cv5nm": failed to list *v1.Secret: secrets "cinder-operator-controller-manager-dockercfg-cv5nm" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Dec 06 10:51:52 crc kubenswrapper[4678]: E1206 10:51:52.919744 4678 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"cinder-operator-controller-manager-dockercfg-cv5nm\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cinder-operator-controller-manager-dockercfg-cv5nm\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.944713 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw"] Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.967965 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb"] Dec 06 10:51:52 crc kubenswrapper[4678]: I1206 10:51:52.968917 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" Dec 06 10:51:52 crc kubenswrapper[4678]: W1206 10:51:52.976141 4678 reflector.go:561] object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-r8dvt": failed to list *v1.Secret: secrets "designate-operator-controller-manager-dockercfg-r8dvt" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Dec 06 10:51:52 crc kubenswrapper[4678]: E1206 10:51:52.976186 4678 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"designate-operator-controller-manager-dockercfg-r8dvt\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"designate-operator-controller-manager-dockercfg-r8dvt\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.002906 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.006427 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg6kw\" (UniqueName: \"kubernetes.io/projected/6669b3a4-9c79-441e-90a2-abceaa89fc58-kube-api-access-dg6kw\") pod \"barbican-operator-controller-manager-7d9dfd778-th2qw\" (UID: \"6669b3a4-9c79-441e-90a2-abceaa89fc58\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.006518 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2jg5\" (UniqueName: \"kubernetes.io/projected/e48c5af7-e126-49e1-9bc0-eec4bd474080-kube-api-access-r2jg5\") pod \"cinder-operator-controller-manager-6c677c69b-zftrw\" (UID: \"e48c5af7-e126-49e1-9bc0-eec4bd474080\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.032813 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.033876 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.038705 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-rtwt2" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.068333 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.096626 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.097834 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.100549 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-744hj" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.107900 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqng5\" (UniqueName: \"kubernetes.io/projected/807c6a8a-a12e-4ac6-8d6a-01f54876d4dd-kube-api-access-sqng5\") pod \"designate-operator-controller-manager-697fb699cf-xqpcb\" (UID: \"807c6a8a-a12e-4ac6-8d6a-01f54876d4dd\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.108127 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg6kw\" (UniqueName: \"kubernetes.io/projected/6669b3a4-9c79-441e-90a2-abceaa89fc58-kube-api-access-dg6kw\") pod \"barbican-operator-controller-manager-7d9dfd778-th2qw\" (UID: \"6669b3a4-9c79-441e-90a2-abceaa89fc58\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.108266 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2jg5\" (UniqueName: \"kubernetes.io/projected/e48c5af7-e126-49e1-9bc0-eec4bd474080-kube-api-access-r2jg5\") pod \"cinder-operator-controller-manager-6c677c69b-zftrw\" (UID: \"e48c5af7-e126-49e1-9bc0-eec4bd474080\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.108371 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.109609 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.115750 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-h6bq7" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.201692 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2jg5\" (UniqueName: \"kubernetes.io/projected/e48c5af7-e126-49e1-9bc0-eec4bd474080-kube-api-access-r2jg5\") pod \"cinder-operator-controller-manager-6c677c69b-zftrw\" (UID: \"e48c5af7-e126-49e1-9bc0-eec4bd474080\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.210988 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xgm5\" (UniqueName: \"kubernetes.io/projected/961960ba-f29c-40e1-81c4-5f8b43456dca-kube-api-access-2xgm5\") pod \"glance-operator-controller-manager-5697bb5779-jpgnt\" (UID: \"961960ba-f29c-40e1-81c4-5f8b43456dca\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.211073 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppmrg\" (UniqueName: \"kubernetes.io/projected/c58c8710-6867-43d2-8fb4-ea96041e2422-kube-api-access-ppmrg\") pod \"heat-operator-controller-manager-5f64f6f8bb-ktmz7\" (UID: \"c58c8710-6867-43d2-8fb4-ea96041e2422\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.211097 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvjb4\" (UniqueName: \"kubernetes.io/projected/b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef-kube-api-access-jvjb4\") pod \"horizon-operator-controller-manager-68c6d99b8f-9kgbd\" (UID: \"b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.211186 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqng5\" (UniqueName: \"kubernetes.io/projected/807c6a8a-a12e-4ac6-8d6a-01f54876d4dd-kube-api-access-sqng5\") pod \"designate-operator-controller-manager-697fb699cf-xqpcb\" (UID: \"807c6a8a-a12e-4ac6-8d6a-01f54876d4dd\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.229807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg6kw\" (UniqueName: \"kubernetes.io/projected/6669b3a4-9c79-441e-90a2-abceaa89fc58-kube-api-access-dg6kw\") pod \"barbican-operator-controller-manager-7d9dfd778-th2qw\" (UID: \"6669b3a4-9c79-441e-90a2-abceaa89fc58\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.232660 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.262545 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.264689 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.287373 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.287607 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-78fn7" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.301879 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.311589 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqng5\" (UniqueName: \"kubernetes.io/projected/807c6a8a-a12e-4ac6-8d6a-01f54876d4dd-kube-api-access-sqng5\") pod \"designate-operator-controller-manager-697fb699cf-xqpcb\" (UID: \"807c6a8a-a12e-4ac6-8d6a-01f54876d4dd\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.312323 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xgm5\" (UniqueName: \"kubernetes.io/projected/961960ba-f29c-40e1-81c4-5f8b43456dca-kube-api-access-2xgm5\") pod \"glance-operator-controller-manager-5697bb5779-jpgnt\" (UID: \"961960ba-f29c-40e1-81c4-5f8b43456dca\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.312356 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppmrg\" (UniqueName: \"kubernetes.io/projected/c58c8710-6867-43d2-8fb4-ea96041e2422-kube-api-access-ppmrg\") pod \"heat-operator-controller-manager-5f64f6f8bb-ktmz7\" (UID: \"c58c8710-6867-43d2-8fb4-ea96041e2422\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.312376 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvjb4\" (UniqueName: \"kubernetes.io/projected/b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef-kube-api-access-jvjb4\") pod \"horizon-operator-controller-manager-68c6d99b8f-9kgbd\" (UID: \"b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.313557 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.328895 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-ssssw"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.330573 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.345691 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-ssssw"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.359258 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.360671 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.366055 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-lpjrx" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.367559 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rfgxt"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.368892 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.371476 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-c5tv2" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.407836 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.415207 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltkm6\" (UniqueName: \"kubernetes.io/projected/7bc20ad2-7338-4943-ae21-1ad8a0fd947e-kube-api-access-ltkm6\") pod \"ironic-operator-controller-manager-967d97867-ssssw\" (UID: \"7bc20ad2-7338-4943-ae21-1ad8a0fd947e\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.415293 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lk45\" (UniqueName: \"kubernetes.io/projected/d272c78c-4c56-4faa-a421-ede41c4b2307-kube-api-access-9lk45\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.415325 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.436755 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfgxt"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.453943 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.463760 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xgm5\" (UniqueName: \"kubernetes.io/projected/961960ba-f29c-40e1-81c4-5f8b43456dca-kube-api-access-2xgm5\") pod \"glance-operator-controller-manager-5697bb5779-jpgnt\" (UID: \"961960ba-f29c-40e1-81c4-5f8b43456dca\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.469319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppmrg\" (UniqueName: \"kubernetes.io/projected/c58c8710-6867-43d2-8fb4-ea96041e2422-kube-api-access-ppmrg\") pod \"heat-operator-controller-manager-5f64f6f8bb-ktmz7\" (UID: \"c58c8710-6867-43d2-8fb4-ea96041e2422\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.470670 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvjb4\" (UniqueName: \"kubernetes.io/projected/b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef-kube-api-access-jvjb4\") pod \"horizon-operator-controller-manager-68c6d99b8f-9kgbd\" (UID: \"b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.471883 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.472917 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.476957 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6lrps" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.516653 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-utilities\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.516712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.516747 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltkm6\" (UniqueName: \"kubernetes.io/projected/7bc20ad2-7338-4943-ae21-1ad8a0fd947e-kube-api-access-ltkm6\") pod \"ironic-operator-controller-manager-967d97867-ssssw\" (UID: \"7bc20ad2-7338-4943-ae21-1ad8a0fd947e\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.516791 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cm4j\" (UniqueName: \"kubernetes.io/projected/106042db-c255-463d-8ea0-60bb39830cb8-kube-api-access-8cm4j\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.516833 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jlzk\" (UniqueName: \"kubernetes.io/projected/78fc5203-6af1-49e0-a485-ba9ef0a0b658-kube-api-access-5jlzk\") pod \"keystone-operator-controller-manager-7765d96ddf-p5bxv\" (UID: \"78fc5203-6af1-49e0-a485-ba9ef0a0b658\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.516865 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lk45\" (UniqueName: \"kubernetes.io/projected/d272c78c-4c56-4faa-a421-ede41c4b2307-kube-api-access-9lk45\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.516880 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-catalog-content\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: E1206 10:51:53.517033 4678 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:53 crc kubenswrapper[4678]: E1206 10:51:53.517087 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert podName:d272c78c-4c56-4faa-a421-ede41c4b2307 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:54.017065315 +0000 UTC m=+918.860496754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert") pod "infra-operator-controller-manager-78d48bff9d-9svrt" (UID: "d272c78c-4c56-4faa-a421-ede41c4b2307") : secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.531884 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.593152 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lk45\" (UniqueName: \"kubernetes.io/projected/d272c78c-4c56-4faa-a421-ede41c4b2307-kube-api-access-9lk45\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.603100 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltkm6\" (UniqueName: \"kubernetes.io/projected/7bc20ad2-7338-4943-ae21-1ad8a0fd947e-kube-api-access-ltkm6\") pod \"ironic-operator-controller-manager-967d97867-ssssw\" (UID: \"7bc20ad2-7338-4943-ae21-1ad8a0fd947e\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.619542 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.620846 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.622237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-utilities\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.622293 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cm4j\" (UniqueName: \"kubernetes.io/projected/106042db-c255-463d-8ea0-60bb39830cb8-kube-api-access-8cm4j\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.622336 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jlzk\" (UniqueName: \"kubernetes.io/projected/78fc5203-6af1-49e0-a485-ba9ef0a0b658-kube-api-access-5jlzk\") pod \"keystone-operator-controller-manager-7765d96ddf-p5bxv\" (UID: \"78fc5203-6af1-49e0-a485-ba9ef0a0b658\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.622369 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5g8g\" (UniqueName: \"kubernetes.io/projected/9e2b27dd-544c-4c0a-ba0f-6c0b4243017c-kube-api-access-k5g8g\") pod \"manila-operator-controller-manager-5b5fd79c9c-vdspj\" (UID: \"9e2b27dd-544c-4c0a-ba0f-6c0b4243017c\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.622390 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-catalog-content\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.622822 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-catalog-content\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.623058 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-utilities\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.639192 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.640228 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.647225 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-str8g" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.648232 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-p7ldh" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.667449 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.679824 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.682234 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cm4j\" (UniqueName: \"kubernetes.io/projected/106042db-c255-463d-8ea0-60bb39830cb8-kube-api-access-8cm4j\") pod \"redhat-marketplace-rfgxt\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.685570 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.707433 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jlzk\" (UniqueName: \"kubernetes.io/projected/78fc5203-6af1-49e0-a485-ba9ef0a0b658-kube-api-access-5jlzk\") pod \"keystone-operator-controller-manager-7765d96ddf-p5bxv\" (UID: \"78fc5203-6af1-49e0-a485-ba9ef0a0b658\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.714062 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.714875 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.725956 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5g8g\" (UniqueName: \"kubernetes.io/projected/9e2b27dd-544c-4c0a-ba0f-6c0b4243017c-kube-api-access-k5g8g\") pod \"manila-operator-controller-manager-5b5fd79c9c-vdspj\" (UID: \"9e2b27dd-544c-4c0a-ba0f-6c0b4243017c\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.731878 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8wck\" (UniqueName: \"kubernetes.io/projected/c7983f9e-7a85-4c2c-b689-7213f06a144a-kube-api-access-r8wck\") pod \"mariadb-operator-controller-manager-79c8c4686c-9b7t6\" (UID: \"c7983f9e-7a85-4c2c-b689-7213f06a144a\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.733288 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.744872 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.766029 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.771214 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5g8g\" (UniqueName: \"kubernetes.io/projected/9e2b27dd-544c-4c0a-ba0f-6c0b4243017c-kube-api-access-k5g8g\") pod \"manila-operator-controller-manager-5b5fd79c9c-vdspj\" (UID: \"9e2b27dd-544c-4c0a-ba0f-6c0b4243017c\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.812113 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.813344 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.819992 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.821084 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.830385 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-fd5ld" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.830807 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-7zphz" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.833881 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8wck\" (UniqueName: \"kubernetes.io/projected/c7983f9e-7a85-4c2c-b689-7213f06a144a-kube-api-access-r8wck\") pod \"mariadb-operator-controller-manager-79c8c4686c-9b7t6\" (UID: \"c7983f9e-7a85-4c2c-b689-7213f06a144a\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.834040 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn2rs\" (UniqueName: \"kubernetes.io/projected/d925b372-ccef-42a2-9b2c-de8539f9d070-kube-api-access-vn2rs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-d5h8j\" (UID: \"d925b372-ccef-42a2-9b2c-de8539f9d070\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.847636 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.858370 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.870423 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.888319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8wck\" (UniqueName: \"kubernetes.io/projected/c7983f9e-7a85-4c2c-b689-7213f06a144a-kube-api-access-r8wck\") pod \"mariadb-operator-controller-manager-79c8c4686c-9b7t6\" (UID: \"c7983f9e-7a85-4c2c-b689-7213f06a144a\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.899436 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.900874 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.906132 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-fdvk6" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.914321 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699"] Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.929373 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.938436 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn2rs\" (UniqueName: \"kubernetes.io/projected/d925b372-ccef-42a2-9b2c-de8539f9d070-kube-api-access-vn2rs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-d5h8j\" (UID: \"d925b372-ccef-42a2-9b2c-de8539f9d070\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.938533 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brxkq\" (UniqueName: \"kubernetes.io/projected/6424c424-5901-4d42-a077-8388ef0978b8-kube-api-access-brxkq\") pod \"nova-operator-controller-manager-697bc559fc-4kvsf\" (UID: \"6424c424-5901-4d42-a077-8388ef0978b8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.938621 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4xvd\" (UniqueName: \"kubernetes.io/projected/f1b9824d-0c1c-4cef-bb30-3e0f14719c6d-kube-api-access-f4xvd\") pod \"octavia-operator-controller-manager-998648c74-fpj6x\" (UID: \"f1b9824d-0c1c-4cef-bb30-3e0f14719c6d\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.947016 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-pv5rp" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.964386 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 06 10:51:53 crc kubenswrapper[4678]: I1206 10:51:53.973635 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.005601 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn2rs\" (UniqueName: \"kubernetes.io/projected/d925b372-ccef-42a2-9b2c-de8539f9d070-kube-api-access-vn2rs\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-d5h8j\" (UID: \"d925b372-ccef-42a2-9b2c-de8539f9d070\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.006037 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.043163 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4xvd\" (UniqueName: \"kubernetes.io/projected/f1b9824d-0c1c-4cef-bb30-3e0f14719c6d-kube-api-access-f4xvd\") pod \"octavia-operator-controller-manager-998648c74-fpj6x\" (UID: \"f1b9824d-0c1c-4cef-bb30-3e0f14719c6d\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.079601 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgdns\" (UniqueName: \"kubernetes.io/projected/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-kube-api-access-vgdns\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.079654 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74hfp\" (UniqueName: \"kubernetes.io/projected/ed890ee4-e09b-40cd-9bce-f21d8676a2c7-kube-api-access-74hfp\") pod \"ovn-operator-controller-manager-b6456fdb6-mrqtn\" (UID: \"ed890ee4-e09b-40cd-9bce-f21d8676a2c7\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.079712 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.079743 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brxkq\" (UniqueName: \"kubernetes.io/projected/6424c424-5901-4d42-a077-8388ef0978b8-kube-api-access-brxkq\") pod \"nova-operator-controller-manager-697bc559fc-4kvsf\" (UID: \"6424c424-5901-4d42-a077-8388ef0978b8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.079790 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.080022 4678 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.080081 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert podName:d272c78c-4c56-4faa-a421-ede41c4b2307 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:55.080061875 +0000 UTC m=+919.923493314 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert") pod "infra-operator-controller-manager-78d48bff9d-9svrt" (UID: "d272c78c-4c56-4faa-a421-ede41c4b2307") : secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.087014 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-cv5nm" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.087245 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.093956 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.100004 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.112712 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4xvd\" (UniqueName: \"kubernetes.io/projected/f1b9824d-0c1c-4cef-bb30-3e0f14719c6d-kube-api-access-f4xvd\") pod \"octavia-operator-controller-manager-998648c74-fpj6x\" (UID: \"f1b9824d-0c1c-4cef-bb30-3e0f14719c6d\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.133142 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brxkq\" (UniqueName: \"kubernetes.io/projected/6424c424-5901-4d42-a077-8388ef0978b8-kube-api-access-brxkq\") pod \"nova-operator-controller-manager-697bc559fc-4kvsf\" (UID: \"6424c424-5901-4d42-a077-8388ef0978b8\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.133528 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.141573 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.142503 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.142999 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.157673 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-vdlw5" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.160556 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.167153 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-tqxvl" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.167307 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-r8dvt" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.171293 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.162067 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.185288 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgdns\" (UniqueName: \"kubernetes.io/projected/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-kube-api-access-vgdns\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.185576 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74hfp\" (UniqueName: \"kubernetes.io/projected/ed890ee4-e09b-40cd-9bce-f21d8676a2c7-kube-api-access-74hfp\") pod \"ovn-operator-controller-manager-b6456fdb6-mrqtn\" (UID: \"ed890ee4-e09b-40cd-9bce-f21d8676a2c7\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.185695 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.185930 4678 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.186053 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert podName:20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f nodeName:}" failed. No retries permitted until 2025-12-06 10:51:54.6860319 +0000 UTC m=+919.529463339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fs2699" (UID: "20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.201614 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.239319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgdns\" (UniqueName: \"kubernetes.io/projected/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-kube-api-access-vgdns\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.240465 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74hfp\" (UniqueName: \"kubernetes.io/projected/ed890ee4-e09b-40cd-9bce-f21d8676a2c7-kube-api-access-74hfp\") pod \"ovn-operator-controller-manager-b6456fdb6-mrqtn\" (UID: \"ed890ee4-e09b-40cd-9bce-f21d8676a2c7\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.274520 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.284676 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.287068 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmttl\" (UniqueName: \"kubernetes.io/projected/b0e3bc2c-84eb-4407-ae9b-101a70b9d555-kube-api-access-fmttl\") pod \"swift-operator-controller-manager-9d58d64bc-pknvj\" (UID: \"b0e3bc2c-84eb-4407-ae9b-101a70b9d555\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.287168 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvttp\" (UniqueName: \"kubernetes.io/projected/f1447b71-f9c7-4969-9d8a-d4b8b1f099bb-kube-api-access-zvttp\") pod \"placement-operator-controller-manager-78f8948974-jdg4h\" (UID: \"f1447b71-f9c7-4969-9d8a-d4b8b1f099bb\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.287627 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.325882 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.341523 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-69g7j" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.367210 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.390868 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvttp\" (UniqueName: \"kubernetes.io/projected/f1447b71-f9c7-4969-9d8a-d4b8b1f099bb-kube-api-access-zvttp\") pod \"placement-operator-controller-manager-78f8948974-jdg4h\" (UID: \"f1447b71-f9c7-4969-9d8a-d4b8b1f099bb\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.391373 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmttl\" (UniqueName: \"kubernetes.io/projected/b0e3bc2c-84eb-4407-ae9b-101a70b9d555-kube-api-access-fmttl\") pod \"swift-operator-controller-manager-9d58d64bc-pknvj\" (UID: \"b0e3bc2c-84eb-4407-ae9b-101a70b9d555\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.397063 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.398722 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.411356 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-lzfsr" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.452446 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvttp\" (UniqueName: \"kubernetes.io/projected/f1447b71-f9c7-4969-9d8a-d4b8b1f099bb-kube-api-access-zvttp\") pod \"placement-operator-controller-manager-78f8948974-jdg4h\" (UID: \"f1447b71-f9c7-4969-9d8a-d4b8b1f099bb\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.460238 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmttl\" (UniqueName: \"kubernetes.io/projected/b0e3bc2c-84eb-4407-ae9b-101a70b9d555-kube-api-access-fmttl\") pod \"swift-operator-controller-manager-9d58d64bc-pknvj\" (UID: \"b0e3bc2c-84eb-4407-ae9b-101a70b9d555\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.471519 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.472701 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.475857 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-kt74p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.493378 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82qzf\" (UniqueName: \"kubernetes.io/projected/b458a451-d4c7-4cb3-b7af-90a12424be9b-kube-api-access-82qzf\") pod \"telemetry-operator-controller-manager-58d5ff84df-m4mrc\" (UID: \"b458a451-d4c7-4cb3-b7af-90a12424be9b\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.493561 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k95n2\" (UniqueName: \"kubernetes.io/projected/9b1bcff8-3a58-4c8a-a971-0374ecb16a9b-kube-api-access-k95n2\") pod \"test-operator-controller-manager-5854674fcc-qj7fr\" (UID: \"9b1bcff8-3a58-4c8a-a971-0374ecb16a9b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.500868 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.507710 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.527641 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.542590 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.598236 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k95n2\" (UniqueName: \"kubernetes.io/projected/9b1bcff8-3a58-4c8a-a971-0374ecb16a9b-kube-api-access-k95n2\") pod \"test-operator-controller-manager-5854674fcc-qj7fr\" (UID: \"9b1bcff8-3a58-4c8a-a971-0374ecb16a9b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.598347 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82qzf\" (UniqueName: \"kubernetes.io/projected/b458a451-d4c7-4cb3-b7af-90a12424be9b-kube-api-access-82qzf\") pod \"telemetry-operator-controller-manager-58d5ff84df-m4mrc\" (UID: \"b458a451-d4c7-4cb3-b7af-90a12424be9b\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.598431 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phxgp\" (UniqueName: \"kubernetes.io/projected/90b306bd-29a2-494d-8399-75d95e9b8597-kube-api-access-phxgp\") pod \"watcher-operator-controller-manager-667bd8d554-wl7ww\" (UID: \"90b306bd-29a2-494d-8399-75d95e9b8597\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.656099 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82qzf\" (UniqueName: \"kubernetes.io/projected/b458a451-d4c7-4cb3-b7af-90a12424be9b-kube-api-access-82qzf\") pod \"telemetry-operator-controller-manager-58d5ff84df-m4mrc\" (UID: \"b458a451-d4c7-4cb3-b7af-90a12424be9b\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.666533 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.667530 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.681807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k95n2\" (UniqueName: \"kubernetes.io/projected/9b1bcff8-3a58-4c8a-a971-0374ecb16a9b-kube-api-access-k95n2\") pod \"test-operator-controller-manager-5854674fcc-qj7fr\" (UID: \"9b1bcff8-3a58-4c8a-a971-0374ecb16a9b\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.682233 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.682407 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.682532 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-v6vnh" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.706982 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phxgp\" (UniqueName: \"kubernetes.io/projected/90b306bd-29a2-494d-8399-75d95e9b8597-kube-api-access-phxgp\") pod \"watcher-operator-controller-manager-667bd8d554-wl7ww\" (UID: \"90b306bd-29a2-494d-8399-75d95e9b8597\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.707195 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.707657 4678 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.707731 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert podName:20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f nodeName:}" failed. No retries permitted until 2025-12-06 10:51:55.7077078 +0000 UTC m=+920.551139239 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fs2699" (UID: "20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.721206 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.774546 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phxgp\" (UniqueName: \"kubernetes.io/projected/90b306bd-29a2-494d-8399-75d95e9b8597-kube-api-access-phxgp\") pod \"watcher-operator-controller-manager-667bd8d554-wl7ww\" (UID: \"90b306bd-29a2-494d-8399-75d95e9b8597\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.815930 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.815985 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.816152 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.816211 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpgd8\" (UniqueName: \"kubernetes.io/projected/67b34972-6c66-47d8-885e-38d2a55a4fd9-kube-api-access-jpgd8\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.836807 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.877876 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.892189 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-ssssw"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.906867 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.917224 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf"] Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.917811 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpgd8\" (UniqueName: \"kubernetes.io/projected/67b34972-6c66-47d8-885e-38d2a55a4fd9-kube-api-access-jpgd8\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.917861 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.917951 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.918082 4678 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.918133 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:55.418120243 +0000 UTC m=+920.261551682 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "metrics-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.918222 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.918405 4678 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: E1206 10:51:54.918430 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:55.418423332 +0000 UTC m=+920.261854761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "webhook-server-cert" not found Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.922024 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-4ltnl" Dec 06 10:51:54 crc kubenswrapper[4678]: W1206 10:51:54.971401 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6669b3a4_9c79_441e_90a2_abceaa89fc58.slice/crio-441f1d8fd16e1b5ef9c9b774574eb76a314d043b3044e5d950b80c00521b61a0 WatchSource:0}: Error finding container 441f1d8fd16e1b5ef9c9b774574eb76a314d043b3044e5d950b80c00521b61a0: Status 404 returned error can't find the container with id 441f1d8fd16e1b5ef9c9b774574eb76a314d043b3044e5d950b80c00521b61a0 Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.973569 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpgd8\" (UniqueName: \"kubernetes.io/projected/67b34972-6c66-47d8-885e-38d2a55a4fd9-kube-api-access-jpgd8\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:54 crc kubenswrapper[4678]: I1206 10:51:54.982970 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf"] Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.027288 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph69z\" (UniqueName: \"kubernetes.io/projected/c232b1db-4726-4697-b59f-f80c1d9d2d4d-kube-api-access-ph69z\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wtgcf\" (UID: \"c232b1db-4726-4697-b59f-f80c1d9d2d4d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.125601 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt"] Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.128865 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.129149 4678 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.129318 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert podName:d272c78c-4c56-4faa-a421-ede41c4b2307 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:57.129254158 +0000 UTC m=+921.972685597 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert") pod "infra-operator-controller-manager-78d48bff9d-9svrt" (UID: "d272c78c-4c56-4faa-a421-ede41c4b2307") : secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.153186 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph69z\" (UniqueName: \"kubernetes.io/projected/c232b1db-4726-4697-b59f-f80c1d9d2d4d-kube-api-access-ph69z\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wtgcf\" (UID: \"c232b1db-4726-4697-b59f-f80c1d9d2d4d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.183530 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph69z\" (UniqueName: \"kubernetes.io/projected/c232b1db-4726-4697-b59f-f80c1d9d2d4d-kube-api-access-ph69z\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wtgcf\" (UID: \"c232b1db-4726-4697-b59f-f80c1d9d2d4d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.207275 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv"] Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.365321 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.448178 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7"] Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.467443 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.467596 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.467687 4678 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.467795 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:56.467764532 +0000 UTC m=+921.311195971 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "metrics-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.467802 4678 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.467868 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:56.467846675 +0000 UTC m=+921.311278194 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "webhook-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.754715 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" event={"ID":"6669b3a4-9c79-441e-90a2-abceaa89fc58","Type":"ContainerStarted","Data":"441f1d8fd16e1b5ef9c9b774574eb76a314d043b3044e5d950b80c00521b61a0"} Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.767878 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" event={"ID":"c58c8710-6867-43d2-8fb4-ea96041e2422","Type":"ContainerStarted","Data":"11bd3bc11e3afb6a8a825a04af7e2d30084d2268c01861c4ba8da68fa6b999d3"} Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.788525 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" event={"ID":"78fc5203-6af1-49e0-a485-ba9ef0a0b658","Type":"ContainerStarted","Data":"01668c1c7e4e545ace721f604af888a1ab921ca0bdeed8e33e26cba9b709910a"} Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.789610 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.789824 4678 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: E1206 10:51:55.789932 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert podName:20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f nodeName:}" failed. No retries permitted until 2025-12-06 10:51:57.789914972 +0000 UTC m=+922.633346411 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fs2699" (UID: "20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.795061 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" event={"ID":"961960ba-f29c-40e1-81c4-5f8b43456dca","Type":"ContainerStarted","Data":"fd4b834e94cddc6702e431eeb2840d797cbc6b9a2761ae3c6ab0d91d5f5b599a"} Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.796164 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" event={"ID":"7bc20ad2-7338-4943-ae21-1ad8a0fd947e","Type":"ContainerStarted","Data":"8c4224329be0373d35b0a90126c3a1ae966bad29483d24eb8ca4262b23ef4dd3"} Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.976628 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd"] Dec 06 10:51:55 crc kubenswrapper[4678]: I1206 10:51:55.983958 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.029299 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6"] Dec 06 10:51:56 crc kubenswrapper[4678]: W1206 10:51:56.080765 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1b9824d_0c1c_4cef_bb30_3e0f14719c6d.slice/crio-4f20637ed0884c554137d7d09fc5bce3772608443d42cbf9f5376571eff4ea0f WatchSource:0}: Error finding container 4f20637ed0884c554137d7d09fc5bce3772608443d42cbf9f5376571eff4ea0f: Status 404 returned error can't find the container with id 4f20637ed0884c554137d7d09fc5bce3772608443d42cbf9f5376571eff4ea0f Dec 06 10:51:56 crc kubenswrapper[4678]: W1206 10:51:56.094030 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode48c5af7_e126_49e1_9bc0_eec4bd474080.slice/crio-37881097d3ab5b553add02a31584482beb20e33a974132fc5124b741bb639e31 WatchSource:0}: Error finding container 37881097d3ab5b553add02a31584482beb20e33a974132fc5124b741bb639e31: Status 404 returned error can't find the container with id 37881097d3ab5b553add02a31584482beb20e33a974132fc5124b741bb639e31 Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.098060 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.145223 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.152939 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.156218 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfgxt"] Dec 06 10:51:56 crc kubenswrapper[4678]: W1206 10:51:56.176182 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod106042db_c255_463d_8ea0_60bb39830cb8.slice/crio-c7dc7427d1fc98ed2f022eda7132b1d20c04efcb0198a3d79171c76983d8a2c1 WatchSource:0}: Error finding container c7dc7427d1fc98ed2f022eda7132b1d20c04efcb0198a3d79171c76983d8a2c1: Status 404 returned error can't find the container with id c7dc7427d1fc98ed2f022eda7132b1d20c04efcb0198a3d79171c76983d8a2c1 Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.448527 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.525297 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.525404 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:56 crc kubenswrapper[4678]: E1206 10:51:56.525701 4678 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 10:51:56 crc kubenswrapper[4678]: E1206 10:51:56.525801 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:58.525778123 +0000 UTC m=+923.369209562 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "webhook-server-cert" not found Dec 06 10:51:56 crc kubenswrapper[4678]: E1206 10:51:56.525812 4678 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 10:51:56 crc kubenswrapper[4678]: E1206 10:51:56.525908 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:51:58.525876216 +0000 UTC m=+923.369307655 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "metrics-server-cert" not found Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.686061 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.724278 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.765183 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.812709 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.812774 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.897172 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" event={"ID":"b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef","Type":"ContainerStarted","Data":"5ab8f7eacd64e286af7bcfdb0123222a0febb3c03d5cf6a78ef129eb35dd3302"} Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.960434 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc"] Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.964356 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfgxt" event={"ID":"106042db-c255-463d-8ea0-60bb39830cb8","Type":"ContainerStarted","Data":"c7dc7427d1fc98ed2f022eda7132b1d20c04efcb0198a3d79171c76983d8a2c1"} Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.982147 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" event={"ID":"6424c424-5901-4d42-a077-8388ef0978b8","Type":"ContainerStarted","Data":"02b465f40644124022a1f52d76be885c086154bee630ce88dd7fed894b1e17b7"} Dec 06 10:51:56 crc kubenswrapper[4678]: I1206 10:51:56.996082 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" event={"ID":"f1b9824d-0c1c-4cef-bb30-3e0f14719c6d","Type":"ContainerStarted","Data":"4f20637ed0884c554137d7d09fc5bce3772608443d42cbf9f5376571eff4ea0f"} Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.000827 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" event={"ID":"ed890ee4-e09b-40cd-9bce-f21d8676a2c7","Type":"ContainerStarted","Data":"392be15049b29d90255618672557ef621e596393e387a74797e08fb4a9f5ea39"} Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.013603 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" event={"ID":"9e2b27dd-544c-4c0a-ba0f-6c0b4243017c","Type":"ContainerStarted","Data":"80a720e83ff8e64a822f7fe4f81a431ae51bb45a929048feb0b04cbf1d854b06"} Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.050259 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" event={"ID":"c7983f9e-7a85-4c2c-b689-7213f06a144a","Type":"ContainerStarted","Data":"6353d483f5a24466da2a14258fa00113d9810c47f95a9e36f188f4d0bff30cbc"} Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.050982 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj"] Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.058348 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" event={"ID":"e48c5af7-e126-49e1-9bc0-eec4bd474080","Type":"ContainerStarted","Data":"37881097d3ab5b553add02a31584482beb20e33a974132fc5124b741bb639e31"} Dec 06 10:51:57 crc kubenswrapper[4678]: E1206 10:51:57.059004 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-phxgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-wl7ww_openstack-operators(90b306bd-29a2-494d-8399-75d95e9b8597): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.068686 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww"] Dec 06 10:51:57 crc kubenswrapper[4678]: E1206 10:51:57.082049 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-phxgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-wl7ww_openstack-operators(90b306bd-29a2-494d-8399-75d95e9b8597): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 06 10:51:57 crc kubenswrapper[4678]: E1206 10:51:57.083818 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" podUID="90b306bd-29a2-494d-8399-75d95e9b8597" Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.153388 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:51:57 crc kubenswrapper[4678]: E1206 10:51:57.153599 4678 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:57 crc kubenswrapper[4678]: E1206 10:51:57.153661 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert podName:d272c78c-4c56-4faa-a421-ede41c4b2307 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:01.153639875 +0000 UTC m=+925.997071314 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert") pod "infra-operator-controller-manager-78d48bff9d-9svrt" (UID: "d272c78c-4c56-4faa-a421-ede41c4b2307") : secret "infra-operator-webhook-server-cert" not found Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.419440 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.419487 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.568904 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:57 crc kubenswrapper[4678]: I1206 10:51:57.866605 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:51:57 crc kubenswrapper[4678]: E1206 10:51:57.866791 4678 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:57 crc kubenswrapper[4678]: E1206 10:51:57.866853 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert podName:20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f nodeName:}" failed. No retries permitted until 2025-12-06 10:52:01.866832072 +0000 UTC m=+926.710263511 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fs2699" (UID: "20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.110553 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" event={"ID":"b458a451-d4c7-4cb3-b7af-90a12424be9b","Type":"ContainerStarted","Data":"784ea17bf5596410e8f021d1756dc9057d57249214a12f04c2ff0ef253c78c56"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.118535 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" event={"ID":"9b1bcff8-3a58-4c8a-a971-0374ecb16a9b","Type":"ContainerStarted","Data":"a5dab152308283b4eb938b435999c8e431dd1ed54b12c0f3691899d472910ea9"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.136536 4678 generic.go:334] "Generic (PLEG): container finished" podID="106042db-c255-463d-8ea0-60bb39830cb8" containerID="44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d" exitCode=0 Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.136612 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfgxt" event={"ID":"106042db-c255-463d-8ea0-60bb39830cb8","Type":"ContainerDied","Data":"44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.140405 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" event={"ID":"c232b1db-4726-4697-b59f-f80c1d9d2d4d","Type":"ContainerStarted","Data":"b97ff6dd94350e77d607edaf1bd81859b80010b38e509365191ccbdffc594505"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.142307 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" event={"ID":"f1447b71-f9c7-4969-9d8a-d4b8b1f099bb","Type":"ContainerStarted","Data":"b27025291e268ecb95ba69dc5bfe2b517432791489adab72215759a348039f8c"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.158855 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" event={"ID":"807c6a8a-a12e-4ac6-8d6a-01f54876d4dd","Type":"ContainerStarted","Data":"41991487b92ae951ae2736f3155b41be1b41235046c1e6e601623661d995690e"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.162437 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" event={"ID":"90b306bd-29a2-494d-8399-75d95e9b8597","Type":"ContainerStarted","Data":"4845e6e3390ab2910ac3c6c4bb114136af7e6f8da96e1ee2cf7edc23d7dfbb28"} Dec 06 10:51:58 crc kubenswrapper[4678]: E1206 10:51:58.169727 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" podUID="90b306bd-29a2-494d-8399-75d95e9b8597" Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.169741 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" event={"ID":"d925b372-ccef-42a2-9b2c-de8539f9d070","Type":"ContainerStarted","Data":"8211be6dd21223a7faca3a7fc84f164668b0f68a9d08d9af3cd7d22ebf619ae3"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.193916 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" event={"ID":"b0e3bc2c-84eb-4407-ae9b-101a70b9d555","Type":"ContainerStarted","Data":"36c38cd2538e671260b371b200dfe1694d3d63756f37bedd61d41c1aa845cc0a"} Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.396298 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.598541 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:58 crc kubenswrapper[4678]: I1206 10:51:58.598617 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:51:58 crc kubenswrapper[4678]: E1206 10:51:58.598751 4678 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 10:51:58 crc kubenswrapper[4678]: E1206 10:51:58.598832 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:02.598809511 +0000 UTC m=+927.442240950 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "webhook-server-cert" not found Dec 06 10:51:58 crc kubenswrapper[4678]: E1206 10:51:58.600600 4678 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 10:51:58 crc kubenswrapper[4678]: E1206 10:51:58.600639 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:02.600628888 +0000 UTC m=+927.444060317 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "metrics-server-cert" not found Dec 06 10:51:59 crc kubenswrapper[4678]: I1206 10:51:59.088706 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nldgn"] Dec 06 10:51:59 crc kubenswrapper[4678]: E1206 10:51:59.260572 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" podUID="90b306bd-29a2-494d-8399-75d95e9b8597" Dec 06 10:51:59 crc kubenswrapper[4678]: I1206 10:51:59.505708 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:51:59 crc kubenswrapper[4678]: I1206 10:51:59.505770 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:52:00 crc kubenswrapper[4678]: I1206 10:52:00.235855 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nldgn" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="registry-server" containerID="cri-o://76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4" gracePeriod=2 Dec 06 10:52:01 crc kubenswrapper[4678]: I1206 10:52:01.247171 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:52:01 crc kubenswrapper[4678]: E1206 10:52:01.247584 4678 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 06 10:52:01 crc kubenswrapper[4678]: E1206 10:52:01.247738 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert podName:d272c78c-4c56-4faa-a421-ede41c4b2307 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:09.247691044 +0000 UTC m=+934.091122483 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert") pod "infra-operator-controller-manager-78d48bff9d-9svrt" (UID: "d272c78c-4c56-4faa-a421-ede41c4b2307") : secret "infra-operator-webhook-server-cert" not found Dec 06 10:52:01 crc kubenswrapper[4678]: I1206 10:52:01.960340 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:52:01 crc kubenswrapper[4678]: E1206 10:52:01.960604 4678 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:52:01 crc kubenswrapper[4678]: E1206 10:52:01.960729 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert podName:20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f nodeName:}" failed. No retries permitted until 2025-12-06 10:52:09.960703105 +0000 UTC m=+934.804134544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert") pod "openstack-baremetal-operator-controller-manager-84b575879fs2699" (UID: "20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 06 10:52:02 crc kubenswrapper[4678]: I1206 10:52:02.270427 4678 generic.go:334] "Generic (PLEG): container finished" podID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerID="76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4" exitCode=0 Dec 06 10:52:02 crc kubenswrapper[4678]: I1206 10:52:02.270524 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldgn" event={"ID":"f7e156c4-11f4-435c-bd15-431daf7c97c1","Type":"ContainerDied","Data":"76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4"} Dec 06 10:52:02 crc kubenswrapper[4678]: I1206 10:52:02.672892 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:02 crc kubenswrapper[4678]: I1206 10:52:02.673462 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:02 crc kubenswrapper[4678]: E1206 10:52:02.673153 4678 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 10:52:02 crc kubenswrapper[4678]: E1206 10:52:02.673627 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:10.673595213 +0000 UTC m=+935.517026652 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "metrics-server-cert" not found Dec 06 10:52:02 crc kubenswrapper[4678]: E1206 10:52:02.673691 4678 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 10:52:02 crc kubenswrapper[4678]: E1206 10:52:02.673773 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:10.673748348 +0000 UTC m=+935.517179897 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "webhook-server-cert" not found Dec 06 10:52:07 crc kubenswrapper[4678]: E1206 10:52:07.420039 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4 is running failed: container process not found" containerID="76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 10:52:07 crc kubenswrapper[4678]: E1206 10:52:07.421089 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4 is running failed: container process not found" containerID="76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 10:52:07 crc kubenswrapper[4678]: E1206 10:52:07.421427 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4 is running failed: container process not found" containerID="76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4" cmd=["grpc_health_probe","-addr=:50051"] Dec 06 10:52:07 crc kubenswrapper[4678]: E1206 10:52:07.421460 4678 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-nldgn" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="registry-server" Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.542470 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.555576 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-utilities\") pod \"f7e156c4-11f4-435c-bd15-431daf7c97c1\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.555729 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-catalog-content\") pod \"f7e156c4-11f4-435c-bd15-431daf7c97c1\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.555921 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzwqp\" (UniqueName: \"kubernetes.io/projected/f7e156c4-11f4-435c-bd15-431daf7c97c1-kube-api-access-qzwqp\") pod \"f7e156c4-11f4-435c-bd15-431daf7c97c1\" (UID: \"f7e156c4-11f4-435c-bd15-431daf7c97c1\") " Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.556423 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-utilities" (OuterVolumeSpecName: "utilities") pod "f7e156c4-11f4-435c-bd15-431daf7c97c1" (UID: "f7e156c4-11f4-435c-bd15-431daf7c97c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.579567 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e156c4-11f4-435c-bd15-431daf7c97c1-kube-api-access-qzwqp" (OuterVolumeSpecName: "kube-api-access-qzwqp") pod "f7e156c4-11f4-435c-bd15-431daf7c97c1" (UID: "f7e156c4-11f4-435c-bd15-431daf7c97c1"). InnerVolumeSpecName "kube-api-access-qzwqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.636941 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7e156c4-11f4-435c-bd15-431daf7c97c1" (UID: "f7e156c4-11f4-435c-bd15-431daf7c97c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.657891 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.657939 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7e156c4-11f4-435c-bd15-431daf7c97c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:52:07 crc kubenswrapper[4678]: I1206 10:52:07.657957 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzwqp\" (UniqueName: \"kubernetes.io/projected/f7e156c4-11f4-435c-bd15-431daf7c97c1-kube-api-access-qzwqp\") on node \"crc\" DevicePath \"\"" Dec 06 10:52:08 crc kubenswrapper[4678]: I1206 10:52:08.311734 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nldgn" event={"ID":"f7e156c4-11f4-435c-bd15-431daf7c97c1","Type":"ContainerDied","Data":"dccbc2d254269a5773c1a28b817d72d99027458fd4d0028e90b49b18e35b8d82"} Dec 06 10:52:08 crc kubenswrapper[4678]: I1206 10:52:08.312109 4678 scope.go:117] "RemoveContainer" containerID="76779d3e8ca56f011c0811c30b86884b665f147e449f85024d0eb82f1611bbb4" Dec 06 10:52:08 crc kubenswrapper[4678]: I1206 10:52:08.311831 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nldgn" Dec 06 10:52:08 crc kubenswrapper[4678]: I1206 10:52:08.352367 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nldgn"] Dec 06 10:52:08 crc kubenswrapper[4678]: I1206 10:52:08.362595 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nldgn"] Dec 06 10:52:09 crc kubenswrapper[4678]: I1206 10:52:09.303074 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:52:09 crc kubenswrapper[4678]: I1206 10:52:09.309529 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d272c78c-4c56-4faa-a421-ede41c4b2307-cert\") pod \"infra-operator-controller-manager-78d48bff9d-9svrt\" (UID: \"d272c78c-4c56-4faa-a421-ede41c4b2307\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:52:09 crc kubenswrapper[4678]: I1206 10:52:09.485261 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" path="/var/lib/kubelet/pods/f7e156c4-11f4-435c-bd15-431daf7c97c1/volumes" Dec 06 10:52:09 crc kubenswrapper[4678]: I1206 10:52:09.543433 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:52:10 crc kubenswrapper[4678]: I1206 10:52:10.013606 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:52:10 crc kubenswrapper[4678]: I1206 10:52:10.025262 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f-cert\") pod \"openstack-baremetal-operator-controller-manager-84b575879fs2699\" (UID: \"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:52:10 crc kubenswrapper[4678]: I1206 10:52:10.204244 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:52:10 crc kubenswrapper[4678]: I1206 10:52:10.728925 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:10 crc kubenswrapper[4678]: I1206 10:52:10.729092 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:10 crc kubenswrapper[4678]: E1206 10:52:10.729228 4678 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 06 10:52:10 crc kubenswrapper[4678]: E1206 10:52:10.729391 4678 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 06 10:52:10 crc kubenswrapper[4678]: E1206 10:52:10.729403 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:26.729333715 +0000 UTC m=+951.572765224 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "webhook-server-cert" not found Dec 06 10:52:10 crc kubenswrapper[4678]: E1206 10:52:10.729524 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs podName:67b34972-6c66-47d8-885e-38d2a55a4fd9 nodeName:}" failed. No retries permitted until 2025-12-06 10:52:26.72947576 +0000 UTC m=+951.572907269 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs") pod "openstack-operator-controller-manager-5dcc4c7f94-xld2p" (UID: "67b34972-6c66-47d8-885e-38d2a55a4fd9") : secret "metrics-server-cert" not found Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.691152 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jtlcj"] Dec 06 10:52:11 crc kubenswrapper[4678]: E1206 10:52:11.691507 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="extract-utilities" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.691521 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="extract-utilities" Dec 06 10:52:11 crc kubenswrapper[4678]: E1206 10:52:11.691535 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="registry-server" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.691541 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="registry-server" Dec 06 10:52:11 crc kubenswrapper[4678]: E1206 10:52:11.691566 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="extract-content" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.691573 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="extract-content" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.691721 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7e156c4-11f4-435c-bd15-431daf7c97c1" containerName="registry-server" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.692694 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.706697 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jtlcj"] Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.747952 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-utilities\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.749224 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4skp\" (UniqueName: \"kubernetes.io/projected/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-kube-api-access-j4skp\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.749358 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-catalog-content\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.851382 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-utilities\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.851745 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4skp\" (UniqueName: \"kubernetes.io/projected/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-kube-api-access-j4skp\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.851903 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-catalog-content\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.852310 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-utilities\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.852724 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-catalog-content\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:11 crc kubenswrapper[4678]: I1206 10:52:11.873426 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4skp\" (UniqueName: \"kubernetes.io/projected/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-kube-api-access-j4skp\") pod \"community-operators-jtlcj\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:12 crc kubenswrapper[4678]: I1206 10:52:12.019384 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:52:12 crc kubenswrapper[4678]: E1206 10:52:12.182834 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168" Dec 06 10:52:12 crc kubenswrapper[4678]: E1206 10:52:12.183619 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f4xvd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-fpj6x_openstack-operators(f1b9824d-0c1c-4cef-bb30-3e0f14719c6d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:13 crc kubenswrapper[4678]: E1206 10:52:13.361720 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a" Dec 06 10:52:13 crc kubenswrapper[4678]: E1206 10:52:13.361959 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k5g8g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-vdspj_openstack-operators(9e2b27dd-544c-4c0a-ba0f-6c0b4243017c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:14 crc kubenswrapper[4678]: E1206 10:52:14.135719 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 06 10:52:14 crc kubenswrapper[4678]: E1206 10:52:14.136558 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k95n2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-qj7fr_openstack-operators(9b1bcff8-3a58-4c8a-a971-0374ecb16a9b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:18 crc kubenswrapper[4678]: E1206 10:52:18.636437 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 06 10:52:18 crc kubenswrapper[4678]: E1206 10:52:18.637090 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-74hfp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-mrqtn_openstack-operators(ed890ee4-e09b-40cd-9bce-f21d8676a2c7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:22 crc kubenswrapper[4678]: E1206 10:52:22.243954 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 06 10:52:22 crc kubenswrapper[4678]: E1206 10:52:22.244809 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2xgm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-jpgnt_openstack-operators(961960ba-f29c-40e1-81c4-5f8b43456dca): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:22 crc kubenswrapper[4678]: E1206 10:52:22.892266 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87" Dec 06 10:52:22 crc kubenswrapper[4678]: E1206 10:52:22.892569 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:5bdb3685be3ddc1efd62e16aaf2fa96ead64315e26d52b1b2a7d8ac01baa1e87,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ltkm6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-967d97867-ssssw_openstack-operators(7bc20ad2-7338-4943-ae21-1ad8a0fd947e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:25 crc kubenswrapper[4678]: E1206 10:52:25.324160 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991" Dec 06 10:52:25 crc kubenswrapper[4678]: E1206 10:52:25.324916 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fmttl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-pknvj_openstack-operators(b0e3bc2c-84eb-4407-ae9b-101a70b9d555): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:26 crc kubenswrapper[4678]: E1206 10:52:26.055704 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a" Dec 06 10:52:26 crc kubenswrapper[4678]: E1206 10:52:26.055945 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:900050d3501c0785b227db34b89883efe68247816e5c7427cacb74f8aa10605a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sqng5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-697fb699cf-xqpcb_openstack-operators(807c6a8a-a12e-4ac6-8d6a-01f54876d4dd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:26 crc kubenswrapper[4678]: I1206 10:52:26.731893 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:26 crc kubenswrapper[4678]: I1206 10:52:26.732436 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:26 crc kubenswrapper[4678]: I1206 10:52:26.739964 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-webhook-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:26 crc kubenswrapper[4678]: I1206 10:52:26.740216 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67b34972-6c66-47d8-885e-38d2a55a4fd9-metrics-certs\") pod \"openstack-operator-controller-manager-5dcc4c7f94-xld2p\" (UID: \"67b34972-6c66-47d8-885e-38d2a55a4fd9\") " pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:26 crc kubenswrapper[4678]: I1206 10:52:26.897025 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:27 crc kubenswrapper[4678]: E1206 10:52:27.700601 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 06 10:52:27 crc kubenswrapper[4678]: E1206 10:52:27.701046 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dg6kw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-th2qw_openstack-operators(6669b3a4-9c79-441e-90a2-abceaa89fc58): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:28 crc kubenswrapper[4678]: E1206 10:52:28.460688 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 06 10:52:28 crc kubenswrapper[4678]: E1206 10:52:28.461637 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ppmrg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-ktmz7_openstack-operators(c58c8710-6867-43d2-8fb4-ea96041e2422): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:29 crc kubenswrapper[4678]: I1206 10:52:29.506066 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:52:29 crc kubenswrapper[4678]: I1206 10:52:29.506165 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:52:29 crc kubenswrapper[4678]: I1206 10:52:29.506238 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:52:29 crc kubenswrapper[4678]: I1206 10:52:29.507182 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a520bd68028a84131872db9f2785f29074c8e74fb3018d3a7e8874376510222f"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 10:52:29 crc kubenswrapper[4678]: I1206 10:52:29.507250 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://a520bd68028a84131872db9f2785f29074c8e74fb3018d3a7e8874376510222f" gracePeriod=600 Dec 06 10:52:30 crc kubenswrapper[4678]: I1206 10:52:30.513209 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="a520bd68028a84131872db9f2785f29074c8e74fb3018d3a7e8874376510222f" exitCode=0 Dec 06 10:52:30 crc kubenswrapper[4678]: I1206 10:52:30.513266 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"a520bd68028a84131872db9f2785f29074c8e74fb3018d3a7e8874376510222f"} Dec 06 10:52:31 crc kubenswrapper[4678]: E1206 10:52:31.890961 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 06 10:52:31 crc kubenswrapper[4678]: E1206 10:52:31.891555 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vn2rs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-d5h8j_openstack-operators(d925b372-ccef-42a2-9b2c-de8539f9d070): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:32 crc kubenswrapper[4678]: E1206 10:52:32.420419 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f" Dec 06 10:52:32 crc kubenswrapper[4678]: E1206 10:52:32.420669 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-82qzf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-m4mrc_openstack-operators(b458a451-d4c7-4cb3-b7af-90a12424be9b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:32 crc kubenswrapper[4678]: E1206 10:52:32.908328 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 06 10:52:32 crc kubenswrapper[4678]: E1206 10:52:32.908752 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvjb4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-9kgbd_openstack-operators(b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:35 crc kubenswrapper[4678]: E1206 10:52:35.903365 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 06 10:52:35 crc kubenswrapper[4678]: E1206 10:52:35.903984 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5jlzk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-p5bxv_openstack-operators(78fc5203-6af1-49e0-a485-ba9ef0a0b658): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:36 crc kubenswrapper[4678]: E1206 10:52:36.555875 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8" Dec 06 10:52:36 crc kubenswrapper[4678]: E1206 10:52:36.556016 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-phxgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-wl7ww_openstack-operators(90b306bd-29a2-494d-8399-75d95e9b8597): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:39 crc kubenswrapper[4678]: E1206 10:52:39.699202 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 06 10:52:39 crc kubenswrapper[4678]: I1206 10:52:39.701412 4678 scope.go:117] "RemoveContainer" containerID="254bc4b474591f0e3849f86c4198a21cd58c274756c497618b470048e2028c07" Dec 06 10:52:39 crc kubenswrapper[4678]: E1206 10:52:39.701547 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ph69z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-wtgcf_openstack-operators(c232b1db-4726-4697-b59f-f80c1d9d2d4d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:52:39 crc kubenswrapper[4678]: E1206 10:52:39.703558 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" podUID="c232b1db-4726-4697-b59f-f80c1d9d2d4d" Dec 06 10:52:40 crc kubenswrapper[4678]: I1206 10:52:40.090617 4678 scope.go:117] "RemoveContainer" containerID="83243b8c01c294af1cd8ae3458e3d6c010e560cd9add00939742aa4303a793e0" Dec 06 10:52:40 crc kubenswrapper[4678]: I1206 10:52:40.338421 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt"] Dec 06 10:52:40 crc kubenswrapper[4678]: I1206 10:52:40.343991 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699"] Dec 06 10:52:40 crc kubenswrapper[4678]: I1206 10:52:40.466409 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jtlcj"] Dec 06 10:52:40 crc kubenswrapper[4678]: I1206 10:52:40.543883 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p"] Dec 06 10:52:40 crc kubenswrapper[4678]: E1206 10:52:40.607066 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" podUID="c232b1db-4726-4697-b59f-f80c1d9d2d4d" Dec 06 10:52:40 crc kubenswrapper[4678]: W1206 10:52:40.895563 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20f3fdfd_d2c8_427c_8371_e5d9cefe0c7f.slice/crio-386d487262a50a0714624d1d0c028d2edb0bd90af2548a9543add622347c50c8 WatchSource:0}: Error finding container 386d487262a50a0714624d1d0c028d2edb0bd90af2548a9543add622347c50c8: Status 404 returned error can't find the container with id 386d487262a50a0714624d1d0c028d2edb0bd90af2548a9543add622347c50c8 Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.068614 4678 scope.go:117] "RemoveContainer" containerID="74cc7443fbc9fd2e3762e8b7e18efbd09edd85649bb12e6ad42344d0b88b6e39" Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.647243 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" event={"ID":"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f","Type":"ContainerStarted","Data":"386d487262a50a0714624d1d0c028d2edb0bd90af2548a9543add622347c50c8"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.654482 4678 generic.go:334] "Generic (PLEG): container finished" podID="106042db-c255-463d-8ea0-60bb39830cb8" containerID="87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5" exitCode=0 Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.654616 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfgxt" event={"ID":"106042db-c255-463d-8ea0-60bb39830cb8","Type":"ContainerDied","Data":"87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.668072 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" event={"ID":"f1447b71-f9c7-4969-9d8a-d4b8b1f099bb","Type":"ContainerStarted","Data":"a1e860b73420914f676c6fe788909609d058c0dd30230364af4aa44f2dea9ba5"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.681076 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" event={"ID":"67b34972-6c66-47d8-885e-38d2a55a4fd9","Type":"ContainerStarted","Data":"57d488f595e717e53398bf3f85a21acbd44996dd56e2f2f3b69ee88c4adee0b9"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.700734 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"996baf59448dd4fd88041c7dc359f0a860305f2ace3dab71db17f1d751ccc5a8"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.707535 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" event={"ID":"d272c78c-4c56-4faa-a421-ede41c4b2307","Type":"ContainerStarted","Data":"85b754432625df1c6000c806a1b1a937d8e6f73477166e7010762d8f1aa0d000"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.728334 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" event={"ID":"e48c5af7-e126-49e1-9bc0-eec4bd474080","Type":"ContainerStarted","Data":"b7e2de6629b314013f98890e9e28a4aed54c93ce6b55b3c7e3becb96089bd7b0"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.734369 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtlcj" event={"ID":"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa","Type":"ContainerStarted","Data":"8389b2eccdfae59652fd13cdb938d392af304fe9cbdecf949a6acdb45702ee4b"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.736010 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" event={"ID":"6424c424-5901-4d42-a077-8388ef0978b8","Type":"ContainerStarted","Data":"7627d09a1daf658a81b6f3026d5ed5c3627b31b9a594a5163fd6f685e075e758"} Dec 06 10:52:41 crc kubenswrapper[4678]: I1206 10:52:41.752771 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" event={"ID":"c7983f9e-7a85-4c2c-b689-7213f06a144a","Type":"ContainerStarted","Data":"31d11c190cf166a44a60f2af8d467e24ff6ce8b6e00fc3152e9cb2f22613b483"} Dec 06 10:52:43 crc kubenswrapper[4678]: I1206 10:52:43.789047 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtlcj" event={"ID":"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa","Type":"ContainerStarted","Data":"fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4"} Dec 06 10:52:44 crc kubenswrapper[4678]: I1206 10:52:44.795271 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" event={"ID":"67b34972-6c66-47d8-885e-38d2a55a4fd9","Type":"ContainerStarted","Data":"33798ce973b189c25a9e5c32410be5dd0dde78186afbc7fa88c2772232fc5f03"} Dec 06 10:52:44 crc kubenswrapper[4678]: I1206 10:52:44.795657 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:52:44 crc kubenswrapper[4678]: I1206 10:52:44.798969 4678 generic.go:334] "Generic (PLEG): container finished" podID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerID="fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4" exitCode=0 Dec 06 10:52:44 crc kubenswrapper[4678]: I1206 10:52:44.799009 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtlcj" event={"ID":"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa","Type":"ContainerDied","Data":"fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4"} Dec 06 10:52:44 crc kubenswrapper[4678]: I1206 10:52:44.825036 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" podStartSLOduration=50.825005154 podStartE2EDuration="50.825005154s" podCreationTimestamp="2025-12-06 10:51:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:52:44.822534907 +0000 UTC m=+969.665966346" watchObservedRunningTime="2025-12-06 10:52:44.825005154 +0000 UTC m=+969.668436603" Dec 06 10:52:47 crc kubenswrapper[4678]: E1206 10:52:47.923441 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" podUID="f1b9824d-0c1c-4cef-bb30-3e0f14719c6d" Dec 06 10:52:47 crc kubenswrapper[4678]: E1206 10:52:47.992577 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" podUID="ed890ee4-e09b-40cd-9bce-f21d8676a2c7" Dec 06 10:52:48 crc kubenswrapper[4678]: E1206 10:52:48.015186 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" podUID="d925b372-ccef-42a2-9b2c-de8539f9d070" Dec 06 10:52:48 crc kubenswrapper[4678]: E1206 10:52:48.054883 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" podUID="c58c8710-6867-43d2-8fb4-ea96041e2422" Dec 06 10:52:48 crc kubenswrapper[4678]: E1206 10:52:48.140460 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" podUID="b0e3bc2c-84eb-4407-ae9b-101a70b9d555" Dec 06 10:52:48 crc kubenswrapper[4678]: E1206 10:52:48.186698 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" podUID="b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef" Dec 06 10:52:48 crc kubenswrapper[4678]: E1206 10:52:48.446466 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" podUID="9e2b27dd-544c-4c0a-ba0f-6c0b4243017c" Dec 06 10:52:48 crc kubenswrapper[4678]: E1206 10:52:48.453314 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" podUID="78fc5203-6af1-49e0-a485-ba9ef0a0b658" Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.850307 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" event={"ID":"f1b9824d-0c1c-4cef-bb30-3e0f14719c6d","Type":"ContainerStarted","Data":"97d1b4d8cfd9e1911dd9b74ea596be44e31b760525f9f7e6386730b7326acf06"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.858064 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" event={"ID":"9e2b27dd-544c-4c0a-ba0f-6c0b4243017c","Type":"ContainerStarted","Data":"d607bb3bbe72562ea8f7c1e991e61240b7548410532a9c476751ca77a48bb7d1"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.863204 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" event={"ID":"d925b372-ccef-42a2-9b2c-de8539f9d070","Type":"ContainerStarted","Data":"92fd8e2ec5273e197bed851b454ebc6392f76f93da5f0603811fd00acd20a0b4"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.869940 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" event={"ID":"d272c78c-4c56-4faa-a421-ede41c4b2307","Type":"ContainerStarted","Data":"abe6e05a344db8c18056feae97ebf1e9d7a499f91a338c90112c7c7905360f4c"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.883785 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" event={"ID":"c58c8710-6867-43d2-8fb4-ea96041e2422","Type":"ContainerStarted","Data":"4e935045afb9ab625c99b922be1c8e01b43571bf33e94d71aa3f742041f3d421"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.892988 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" event={"ID":"ed890ee4-e09b-40cd-9bce-f21d8676a2c7","Type":"ContainerStarted","Data":"5fe8feb6116b22e8b86dc1e1dcad1459070562967ea166fbe001fdae0cef4f72"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.918095 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" event={"ID":"78fc5203-6af1-49e0-a485-ba9ef0a0b658","Type":"ContainerStarted","Data":"386492144dd76f5b5e92f3948ec0e8c1a6967f37d7891a9ba671fe730334da62"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.939578 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" event={"ID":"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f","Type":"ContainerStarted","Data":"754b1e1f06f381b4a839457f6f3592fa4e44f3951c0d3076cf4a7ece89ba7abe"} Dec 06 10:52:48 crc kubenswrapper[4678]: I1206 10:52:48.987206 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfgxt" event={"ID":"106042db-c255-463d-8ea0-60bb39830cb8","Type":"ContainerStarted","Data":"d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a"} Dec 06 10:52:49 crc kubenswrapper[4678]: I1206 10:52:49.021949 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" event={"ID":"b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef","Type":"ContainerStarted","Data":"fb1a5af5af9a99770a498ab89e389d9340d3f4cabbeede3186c66a77621efc13"} Dec 06 10:52:49 crc kubenswrapper[4678]: I1206 10:52:49.042840 4678 generic.go:334] "Generic (PLEG): container finished" podID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerID="3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad" exitCode=0 Dec 06 10:52:49 crc kubenswrapper[4678]: I1206 10:52:49.042961 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtlcj" event={"ID":"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa","Type":"ContainerDied","Data":"3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad"} Dec 06 10:52:49 crc kubenswrapper[4678]: I1206 10:52:49.064885 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" event={"ID":"b0e3bc2c-84eb-4407-ae9b-101a70b9d555","Type":"ContainerStarted","Data":"9fc8e933e01b273a06dcc3486c9eff34c67f4689fcbe17257b06ab6ea633d43d"} Dec 06 10:52:49 crc kubenswrapper[4678]: I1206 10:52:49.150517 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rfgxt" podStartSLOduration=6.984828859 podStartE2EDuration="56.150473711s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:58.151082699 +0000 UTC m=+922.994514138" lastFinishedPulling="2025-12-06 10:52:47.316727551 +0000 UTC m=+972.160158990" observedRunningTime="2025-12-06 10:52:49.147408007 +0000 UTC m=+973.990839446" watchObservedRunningTime="2025-12-06 10:52:49.150473711 +0000 UTC m=+973.993905150" Dec 06 10:52:50 crc kubenswrapper[4678]: E1206 10:52:50.742821 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" podUID="9b1bcff8-3a58-4c8a-a971-0374ecb16a9b" Dec 06 10:52:50 crc kubenswrapper[4678]: E1206 10:52:50.811409 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" podUID="961960ba-f29c-40e1-81c4-5f8b43456dca" Dec 06 10:52:50 crc kubenswrapper[4678]: E1206 10:52:50.811554 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" podUID="7bc20ad2-7338-4943-ae21-1ad8a0fd947e" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.098513 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" event={"ID":"6424c424-5901-4d42-a077-8388ef0978b8","Type":"ContainerStarted","Data":"38d60fbcf7731933c93b0ab46b2dd078c21b60a16252fb3b95037f4a722ff50b"} Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.099277 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.101815 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.108537 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" event={"ID":"f1447b71-f9c7-4969-9d8a-d4b8b1f099bb","Type":"ContainerStarted","Data":"38793f2c8fa3c9fe2f9949cbfde860f568832fe7ab88488775f12c7dcc8a8158"} Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.108832 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.110938 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.112367 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" event={"ID":"961960ba-f29c-40e1-81c4-5f8b43456dca","Type":"ContainerStarted","Data":"6133f33b2da2537288d12cc2c90bed59f5735d54e9f9338a264a5bca66bfbfb3"} Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.113587 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" event={"ID":"c7983f9e-7a85-4c2c-b689-7213f06a144a","Type":"ContainerStarted","Data":"7e89844b1ab6f5c9f4db49aee7d9f25f046d61aef838d7601b27f7fb34ff076b"} Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.113991 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.115569 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" event={"ID":"9b1bcff8-3a58-4c8a-a971-0374ecb16a9b","Type":"ContainerStarted","Data":"c0b726a1192fb20023bd0eb70bd49b0147f71266c5089f418c3858f473ecadcc"} Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.117177 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.117472 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" event={"ID":"7bc20ad2-7338-4943-ae21-1ad8a0fd947e","Type":"ContainerStarted","Data":"7c5f4aaee24eb165b4ceb7ede63ad243362195e7dcda3f3d04e53637e8acb8c6"} Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.118935 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" event={"ID":"e48c5af7-e126-49e1-9bc0-eec4bd474080","Type":"ContainerStarted","Data":"c76415a04e7587221d7ec271a80724d9c01769a38dfff8ec9102bb160582230a"} Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.119574 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.122196 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.130309 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-4kvsf" podStartSLOduration=6.827814207 podStartE2EDuration="58.130288255s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.154179617 +0000 UTC m=+920.997611056" lastFinishedPulling="2025-12-06 10:52:47.456653675 +0000 UTC m=+972.300085104" observedRunningTime="2025-12-06 10:52:51.124973933 +0000 UTC m=+975.968405382" watchObservedRunningTime="2025-12-06 10:52:51.130288255 +0000 UTC m=+975.973719694" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.204222 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9b7t6" podStartSLOduration=6.74526431 podStartE2EDuration="58.20420185s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.04560785 +0000 UTC m=+920.889039289" lastFinishedPulling="2025-12-06 10:52:47.50454539 +0000 UTC m=+972.347976829" observedRunningTime="2025-12-06 10:52:51.156471107 +0000 UTC m=+975.999902546" watchObservedRunningTime="2025-12-06 10:52:51.20420185 +0000 UTC m=+976.047633289" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.239606 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-zftrw" podStartSLOduration=7.84619626 podStartE2EDuration="59.239589695s" podCreationTimestamp="2025-12-06 10:51:52 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.102797459 +0000 UTC m=+920.946228898" lastFinishedPulling="2025-12-06 10:52:47.496190894 +0000 UTC m=+972.339622333" observedRunningTime="2025-12-06 10:52:51.238802951 +0000 UTC m=+976.082234390" watchObservedRunningTime="2025-12-06 10:52:51.239589695 +0000 UTC m=+976.083021134" Dec 06 10:52:51 crc kubenswrapper[4678]: I1206 10:52:51.294618 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-jdg4h" podStartSLOduration=7.751713221 podStartE2EDuration="58.29459697s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.903960126 +0000 UTC m=+921.747391565" lastFinishedPulling="2025-12-06 10:52:47.446843875 +0000 UTC m=+972.290275314" observedRunningTime="2025-12-06 10:52:51.291211876 +0000 UTC m=+976.134643315" watchObservedRunningTime="2025-12-06 10:52:51.29459697 +0000 UTC m=+976.138028409" Dec 06 10:52:51 crc kubenswrapper[4678]: E1206 10:52:51.584419 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" podUID="90b306bd-29a2-494d-8399-75d95e9b8597" Dec 06 10:52:51 crc kubenswrapper[4678]: E1206 10:52:51.602884 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" podUID="b458a451-d4c7-4cb3-b7af-90a12424be9b" Dec 06 10:52:51 crc kubenswrapper[4678]: E1206 10:52:51.606026 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" podUID="6669b3a4-9c79-441e-90a2-abceaa89fc58" Dec 06 10:52:51 crc kubenswrapper[4678]: E1206 10:52:51.609443 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" podUID="807c6a8a-a12e-4ac6-8d6a-01f54876d4dd" Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.139522 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" event={"ID":"20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f","Type":"ContainerStarted","Data":"b93fcd80956891383504facd3045e73b5558a1d2794868f112ac674ff64a2e5a"} Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.140195 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.154927 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" event={"ID":"807c6a8a-a12e-4ac6-8d6a-01f54876d4dd","Type":"ContainerStarted","Data":"e0b76474ab3f3d02950464d09bd4aab4453c758e3b6f3b4c36c6bd6c10d390f0"} Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.158768 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.159872 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" event={"ID":"b458a451-d4c7-4cb3-b7af-90a12424be9b","Type":"ContainerStarted","Data":"e964fdbd8893aeea6f6145f98bffa5194f1c4e495f59dd635991f36edc20a8db"} Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.177028 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-84b575879fs2699" podStartSLOduration=52.725988089 podStartE2EDuration="59.177009993s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:52:40.899194931 +0000 UTC m=+965.742626370" lastFinishedPulling="2025-12-06 10:52:47.350216835 +0000 UTC m=+972.193648274" observedRunningTime="2025-12-06 10:52:52.174693452 +0000 UTC m=+977.018124881" watchObservedRunningTime="2025-12-06 10:52:52.177009993 +0000 UTC m=+977.020441432" Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.191916 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" event={"ID":"90b306bd-29a2-494d-8399-75d95e9b8597","Type":"ContainerStarted","Data":"aadef8f59de5251c7bb163cfd2eb88a46f28afbfd5e56d957c2bd5623e54a3ba"} Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.203927 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" event={"ID":"6669b3a4-9c79-441e-90a2-abceaa89fc58","Type":"ContainerStarted","Data":"36b796d847290a7f2fb65734eced233471c1d18dbfcaf48582e1c4e88d002072"} Dec 06 10:52:52 crc kubenswrapper[4678]: E1206 10:52:52.238648 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" podUID="90b306bd-29a2-494d-8399-75d95e9b8597" Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.247614 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" event={"ID":"d272c78c-4c56-4faa-a421-ede41c4b2307","Type":"ContainerStarted","Data":"de3be1d6b9881e6788adc69a141fa465f5c1d2955f26c601d40d664c503df1f5"} Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.250363 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.271910 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" Dec 06 10:52:52 crc kubenswrapper[4678]: I1206 10:52:52.426286 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-9svrt" podStartSLOduration=52.945640646 podStartE2EDuration="59.426252202s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:52:40.887233894 +0000 UTC m=+965.730665333" lastFinishedPulling="2025-12-06 10:52:47.36784545 +0000 UTC m=+972.211276889" observedRunningTime="2025-12-06 10:52:52.416690938 +0000 UTC m=+977.260122397" watchObservedRunningTime="2025-12-06 10:52:52.426252202 +0000 UTC m=+977.269683641" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.275199 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" event={"ID":"b0e3bc2c-84eb-4407-ae9b-101a70b9d555","Type":"ContainerStarted","Data":"bd6c21df82fcd4238a43dee4c5cfd62efd2cd23a76eb8e2642578a18c726fab0"} Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.276656 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.283045 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" event={"ID":"ed890ee4-e09b-40cd-9bce-f21d8676a2c7","Type":"ContainerStarted","Data":"eea243719185681a8348c9e12a45225922c9de5d4835357dbc52ee134d1cfa6d"} Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.283369 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.288564 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" event={"ID":"78fc5203-6af1-49e0-a485-ba9ef0a0b658","Type":"ContainerStarted","Data":"f2d34ec5d45ece727e3e0de96f27b21f998a02e37f62b74c3baf4c997cf45322"} Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.289018 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.294197 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" event={"ID":"f1b9824d-0c1c-4cef-bb30-3e0f14719c6d","Type":"ContainerStarted","Data":"1353542ffbad1c3aeca5e438f1cac0b889855a594b43422682f6d5689ed61d08"} Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.294673 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.309680 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" event={"ID":"c58c8710-6867-43d2-8fb4-ea96041e2422","Type":"ContainerStarted","Data":"36bad38d507d9b0d451d6dff998b6b12a666bff74466008c6c90a9f90d773cdb"} Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.310433 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.361903 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtlcj" event={"ID":"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa","Type":"ContainerStarted","Data":"393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1"} Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.394785 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" podStartSLOduration=5.32671803 podStartE2EDuration="1m0.394757183s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:57.015185676 +0000 UTC m=+921.858617115" lastFinishedPulling="2025-12-06 10:52:52.083224839 +0000 UTC m=+976.926656268" observedRunningTime="2025-12-06 10:52:53.342889393 +0000 UTC m=+978.186320832" watchObservedRunningTime="2025-12-06 10:52:53.394757183 +0000 UTC m=+978.238188632" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.397234 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" podStartSLOduration=3.808806499 podStartE2EDuration="1m0.397225898s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:55.490152217 +0000 UTC m=+920.333583656" lastFinishedPulling="2025-12-06 10:52:52.078571626 +0000 UTC m=+976.922003055" observedRunningTime="2025-12-06 10:52:53.38846859 +0000 UTC m=+978.231900029" watchObservedRunningTime="2025-12-06 10:52:53.397225898 +0000 UTC m=+978.240657337" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.562235 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" podStartSLOduration=4.585911602 podStartE2EDuration="1m0.562202184s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.102995196 +0000 UTC m=+920.946426635" lastFinishedPulling="2025-12-06 10:52:52.079285778 +0000 UTC m=+976.922717217" observedRunningTime="2025-12-06 10:52:53.44460837 +0000 UTC m=+978.288039819" watchObservedRunningTime="2025-12-06 10:52:53.562202184 +0000 UTC m=+978.405633623" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.568596 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" podStartSLOduration=5.097970314 podStartE2EDuration="1m0.568561479s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.608653182 +0000 UTC m=+921.452084621" lastFinishedPulling="2025-12-06 10:52:52.079244347 +0000 UTC m=+976.922675786" observedRunningTime="2025-12-06 10:52:53.552671342 +0000 UTC m=+978.396102781" watchObservedRunningTime="2025-12-06 10:52:53.568561479 +0000 UTC m=+978.411992918" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.623507 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" podStartSLOduration=3.90726314 podStartE2EDuration="1m0.623470892s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:55.363403727 +0000 UTC m=+920.206835176" lastFinishedPulling="2025-12-06 10:52:52.079611489 +0000 UTC m=+976.923042928" observedRunningTime="2025-12-06 10:52:53.608102751 +0000 UTC m=+978.451534180" watchObservedRunningTime="2025-12-06 10:52:53.623470892 +0000 UTC m=+978.466902331" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.678810 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jtlcj" podStartSLOduration=35.373052042 podStartE2EDuration="42.678782516s" podCreationTimestamp="2025-12-06 10:52:11 +0000 UTC" firstStartedPulling="2025-12-06 10:52:45.244218469 +0000 UTC m=+970.087649948" lastFinishedPulling="2025-12-06 10:52:52.549948983 +0000 UTC m=+977.393380422" observedRunningTime="2025-12-06 10:52:53.661448136 +0000 UTC m=+978.504879575" watchObservedRunningTime="2025-12-06 10:52:53.678782516 +0000 UTC m=+978.522213955" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.738727 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:52:53 crc kubenswrapper[4678]: I1206 10:52:53.738771 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.371041 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" event={"ID":"b458a451-d4c7-4cb3-b7af-90a12424be9b","Type":"ContainerStarted","Data":"bca4b92254234213cd7794071e5acef679ca3f58a700cae2d095f1db74e71dcb"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.371456 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.373324 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" event={"ID":"b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef","Type":"ContainerStarted","Data":"39c5c856c63aca87feb73e443fbf0598a5dcc019eaabdb548ce1f50aa2b60d42"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.373810 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.375916 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" event={"ID":"6669b3a4-9c79-441e-90a2-abceaa89fc58","Type":"ContainerStarted","Data":"6349786297c951714b23da9d10e3a966547effe79497fb7c4263cbc232ab1590"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.376362 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.380355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" event={"ID":"9b1bcff8-3a58-4c8a-a971-0374ecb16a9b","Type":"ContainerStarted","Data":"14320ab50b742d7dff83c69434c878634bfc154477675804cd33d37f31f959c9"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.381032 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.383503 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" event={"ID":"7bc20ad2-7338-4943-ae21-1ad8a0fd947e","Type":"ContainerStarted","Data":"5c9315d886a66dc3af41ab9b580543ac2e2844a5b2177be821a2c897e88d7a2c"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.384233 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.385763 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" event={"ID":"807c6a8a-a12e-4ac6-8d6a-01f54876d4dd","Type":"ContainerStarted","Data":"05811bf2e5a84283e63a8041ed1fae49cb2950c6478b7d4614b86aa276ed9329"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.386182 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.389585 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" event={"ID":"961960ba-f29c-40e1-81c4-5f8b43456dca","Type":"ContainerStarted","Data":"0fdccef9bd52f28fb0c04a0df1670b41dff0a6427b303cf11c0963b7cc09f541"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.390375 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.391947 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" event={"ID":"9e2b27dd-544c-4c0a-ba0f-6c0b4243017c","Type":"ContainerStarted","Data":"cde787c7655e9d906dfbef0f518ed8e7eb01e68706495f2bfa68e670fa715734"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.392085 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.393552 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" event={"ID":"d925b372-ccef-42a2-9b2c-de8539f9d070","Type":"ContainerStarted","Data":"43e91ca47e1919cdaaaf86784d7e89fe9e2864849a0ddfeb97bf42a2ae9565bb"} Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.484968 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" podStartSLOduration=4.667433111 podStartE2EDuration="1m0.484945462s" podCreationTimestamp="2025-12-06 10:51:54 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.845107474 +0000 UTC m=+921.688538913" lastFinishedPulling="2025-12-06 10:52:52.662619825 +0000 UTC m=+977.506051264" observedRunningTime="2025-12-06 10:52:54.479733423 +0000 UTC m=+979.323164862" watchObservedRunningTime="2025-12-06 10:52:54.484945462 +0000 UTC m=+979.328376901" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.488014 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" podStartSLOduration=4.451138924 podStartE2EDuration="1m0.487996086s" podCreationTimestamp="2025-12-06 10:51:54 +0000 UTC" firstStartedPulling="2025-12-06 10:51:57.014922139 +0000 UTC m=+921.858353578" lastFinishedPulling="2025-12-06 10:52:53.051779301 +0000 UTC m=+977.895210740" observedRunningTime="2025-12-06 10:52:54.419716304 +0000 UTC m=+979.263147743" watchObservedRunningTime="2025-12-06 10:52:54.487996086 +0000 UTC m=+979.331427525" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.546537 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" podStartSLOduration=5.011629988 podStartE2EDuration="1m1.546507809s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.030089701 +0000 UTC m=+920.873521140" lastFinishedPulling="2025-12-06 10:52:52.564967532 +0000 UTC m=+977.408398961" observedRunningTime="2025-12-06 10:52:54.545577221 +0000 UTC m=+979.389008660" watchObservedRunningTime="2025-12-06 10:52:54.546507809 +0000 UTC m=+979.389939238" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.644915 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" podStartSLOduration=4.266913374 podStartE2EDuration="1m1.644894564s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:55.197284259 +0000 UTC m=+920.040715698" lastFinishedPulling="2025-12-06 10:52:52.575265449 +0000 UTC m=+977.418696888" observedRunningTime="2025-12-06 10:52:54.627767699 +0000 UTC m=+979.471199138" watchObservedRunningTime="2025-12-06 10:52:54.644894564 +0000 UTC m=+979.488326003" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.718831 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" podStartSLOduration=4.508905713 podStartE2EDuration="1m1.71880554s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:55.379165574 +0000 UTC m=+920.222597013" lastFinishedPulling="2025-12-06 10:52:52.589065401 +0000 UTC m=+977.432496840" observedRunningTime="2025-12-06 10:52:54.673289555 +0000 UTC m=+979.516720994" watchObservedRunningTime="2025-12-06 10:52:54.71880554 +0000 UTC m=+979.562236979" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.767876 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" podStartSLOduration=5.982481608 podStartE2EDuration="1m1.767853682s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.907763976 +0000 UTC m=+921.751195415" lastFinishedPulling="2025-12-06 10:52:52.69313605 +0000 UTC m=+977.536567489" observedRunningTime="2025-12-06 10:52:54.723450662 +0000 UTC m=+979.566882101" watchObservedRunningTime="2025-12-06 10:52:54.767853682 +0000 UTC m=+979.611285121" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.801456 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" podStartSLOduration=6.627502376 podStartE2EDuration="1m2.801434211s" podCreationTimestamp="2025-12-06 10:51:52 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.845064882 +0000 UTC m=+921.688496321" lastFinishedPulling="2025-12-06 10:52:53.018996717 +0000 UTC m=+977.862428156" observedRunningTime="2025-12-06 10:52:54.770729811 +0000 UTC m=+979.614161240" watchObservedRunningTime="2025-12-06 10:52:54.801434211 +0000 UTC m=+979.644865650" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.839419 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" podStartSLOduration=4.850326495 podStartE2EDuration="1m2.839399336s" podCreationTimestamp="2025-12-06 10:51:52 +0000 UTC" firstStartedPulling="2025-12-06 10:51:55.083359033 +0000 UTC m=+919.926790472" lastFinishedPulling="2025-12-06 10:52:53.072431884 +0000 UTC m=+977.915863313" observedRunningTime="2025-12-06 10:52:54.803616399 +0000 UTC m=+979.647047838" watchObservedRunningTime="2025-12-06 10:52:54.839399336 +0000 UTC m=+979.682830775" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.842712 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" podStartSLOduration=5.478487435 podStartE2EDuration="1m1.842686956s" podCreationTimestamp="2025-12-06 10:51:53 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.004601409 +0000 UTC m=+920.848032838" lastFinishedPulling="2025-12-06 10:52:52.36880092 +0000 UTC m=+977.212232359" observedRunningTime="2025-12-06 10:52:54.837404984 +0000 UTC m=+979.680836423" watchObservedRunningTime="2025-12-06 10:52:54.842686956 +0000 UTC m=+979.686118395" Dec 06 10:52:54 crc kubenswrapper[4678]: I1206 10:52:54.918246 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-rfgxt" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="registry-server" probeResult="failure" output=< Dec 06 10:52:54 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 10:52:54 crc kubenswrapper[4678]: > Dec 06 10:52:55 crc kubenswrapper[4678]: I1206 10:52:55.402641 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" Dec 06 10:52:56 crc kubenswrapper[4678]: I1206 10:52:56.409850 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" event={"ID":"c232b1db-4726-4697-b59f-f80c1d9d2d4d","Type":"ContainerStarted","Data":"63e65cac9681c91bb2ce7fc97292f55762650f071074c54e840639001d11457c"} Dec 06 10:52:56 crc kubenswrapper[4678]: I1206 10:52:56.427612 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wtgcf" podStartSLOduration=3.443530556 podStartE2EDuration="1m2.427593497s" podCreationTimestamp="2025-12-06 10:51:54 +0000 UTC" firstStartedPulling="2025-12-06 10:51:56.909020465 +0000 UTC m=+921.752451904" lastFinishedPulling="2025-12-06 10:52:55.893083406 +0000 UTC m=+980.736514845" observedRunningTime="2025-12-06 10:52:56.426946898 +0000 UTC m=+981.270378337" watchObservedRunningTime="2025-12-06 10:52:56.427593497 +0000 UTC m=+981.271024936" Dec 06 10:52:56 crc kubenswrapper[4678]: I1206 10:52:56.903508 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5dcc4c7f94-xld2p" Dec 06 10:53:02 crc kubenswrapper[4678]: I1206 10:53:02.020250 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:53:02 crc kubenswrapper[4678]: I1206 10:53:02.020976 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:53:02 crc kubenswrapper[4678]: I1206 10:53:02.073110 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:53:02 crc kubenswrapper[4678]: I1206 10:53:02.504248 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:53:02 crc kubenswrapper[4678]: I1206 10:53:02.555876 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jtlcj"] Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.458029 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-th2qw" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.671705 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-jpgnt" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.682713 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-ssssw" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.719736 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-ktmz7" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.720883 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-p5bxv" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.779573 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-9kgbd" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.826557 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.886340 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-vdspj" Dec 06 10:53:03 crc kubenswrapper[4678]: I1206 10:53:03.895049 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.176295 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-xqpcb" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.183196 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-fpj6x" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.277928 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-mrqtn" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.287261 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-d5h8j" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.470797 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jtlcj" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="registry-server" containerID="cri-o://393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1" gracePeriod=2 Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.511166 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-pknvj" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.713212 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfgxt"] Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.840771 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-m4mrc" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.872075 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.882625 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qj7fr" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.945922 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4skp\" (UniqueName: \"kubernetes.io/projected/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-kube-api-access-j4skp\") pod \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.946025 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-utilities\") pod \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.946163 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-catalog-content\") pod \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\" (UID: \"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa\") " Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.952945 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-utilities" (OuterVolumeSpecName: "utilities") pod "4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" (UID: "4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:53:04 crc kubenswrapper[4678]: I1206 10:53:04.972125 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-kube-api-access-j4skp" (OuterVolumeSpecName: "kube-api-access-j4skp") pod "4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" (UID: "4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa"). InnerVolumeSpecName "kube-api-access-j4skp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.002931 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" (UID: "4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.048097 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.048134 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.048145 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4skp\" (UniqueName: \"kubernetes.io/projected/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa-kube-api-access-j4skp\") on node \"crc\" DevicePath \"\"" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.486132 4678 generic.go:334] "Generic (PLEG): container finished" podID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerID="393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1" exitCode=0 Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.487421 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rfgxt" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="registry-server" containerID="cri-o://d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a" gracePeriod=2 Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.487573 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jtlcj" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.501944 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtlcj" event={"ID":"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa","Type":"ContainerDied","Data":"393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1"} Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.502004 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jtlcj" event={"ID":"4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa","Type":"ContainerDied","Data":"8389b2eccdfae59652fd13cdb938d392af304fe9cbdecf949a6acdb45702ee4b"} Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.502067 4678 scope.go:117] "RemoveContainer" containerID="393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.536692 4678 scope.go:117] "RemoveContainer" containerID="3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.551895 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jtlcj"] Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.562397 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jtlcj"] Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.564262 4678 scope.go:117] "RemoveContainer" containerID="fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.667777 4678 scope.go:117] "RemoveContainer" containerID="393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1" Dec 06 10:53:05 crc kubenswrapper[4678]: E1206 10:53:05.668686 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1\": container with ID starting with 393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1 not found: ID does not exist" containerID="393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.668723 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1"} err="failed to get container status \"393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1\": rpc error: code = NotFound desc = could not find container \"393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1\": container with ID starting with 393ae92ed3a15b22dcb382ee6a4e610c5a793906433a759b9e02ef3c34a8b2d1 not found: ID does not exist" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.668752 4678 scope.go:117] "RemoveContainer" containerID="3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad" Dec 06 10:53:05 crc kubenswrapper[4678]: E1206 10:53:05.669031 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad\": container with ID starting with 3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad not found: ID does not exist" containerID="3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.669051 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad"} err="failed to get container status \"3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad\": rpc error: code = NotFound desc = could not find container \"3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad\": container with ID starting with 3ba5793b3054307739e88319cb4ed044850ae62005e42c5af2d33714fe0ee7ad not found: ID does not exist" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.669067 4678 scope.go:117] "RemoveContainer" containerID="fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4" Dec 06 10:53:05 crc kubenswrapper[4678]: E1206 10:53:05.669334 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4\": container with ID starting with fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4 not found: ID does not exist" containerID="fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.669355 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4"} err="failed to get container status \"fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4\": rpc error: code = NotFound desc = could not find container \"fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4\": container with ID starting with fa408c479e19dd87637192dbb2817e1af3c66a9d536a51fe89f481de914824b4 not found: ID does not exist" Dec 06 10:53:05 crc kubenswrapper[4678]: I1206 10:53:05.956906 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.076591 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-catalog-content\") pod \"106042db-c255-463d-8ea0-60bb39830cb8\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.077056 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-utilities\") pod \"106042db-c255-463d-8ea0-60bb39830cb8\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.077192 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cm4j\" (UniqueName: \"kubernetes.io/projected/106042db-c255-463d-8ea0-60bb39830cb8-kube-api-access-8cm4j\") pod \"106042db-c255-463d-8ea0-60bb39830cb8\" (UID: \"106042db-c255-463d-8ea0-60bb39830cb8\") " Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.080620 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-utilities" (OuterVolumeSpecName: "utilities") pod "106042db-c255-463d-8ea0-60bb39830cb8" (UID: "106042db-c255-463d-8ea0-60bb39830cb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.088736 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/106042db-c255-463d-8ea0-60bb39830cb8-kube-api-access-8cm4j" (OuterVolumeSpecName: "kube-api-access-8cm4j") pod "106042db-c255-463d-8ea0-60bb39830cb8" (UID: "106042db-c255-463d-8ea0-60bb39830cb8"). InnerVolumeSpecName "kube-api-access-8cm4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.100056 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "106042db-c255-463d-8ea0-60bb39830cb8" (UID: "106042db-c255-463d-8ea0-60bb39830cb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.179212 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.179268 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/106042db-c255-463d-8ea0-60bb39830cb8-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.179281 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cm4j\" (UniqueName: \"kubernetes.io/projected/106042db-c255-463d-8ea0-60bb39830cb8-kube-api-access-8cm4j\") on node \"crc\" DevicePath \"\"" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.499732 4678 generic.go:334] "Generic (PLEG): container finished" podID="106042db-c255-463d-8ea0-60bb39830cb8" containerID="d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a" exitCode=0 Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.499818 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfgxt" event={"ID":"106042db-c255-463d-8ea0-60bb39830cb8","Type":"ContainerDied","Data":"d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a"} Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.499845 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rfgxt" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.499855 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rfgxt" event={"ID":"106042db-c255-463d-8ea0-60bb39830cb8","Type":"ContainerDied","Data":"c7dc7427d1fc98ed2f022eda7132b1d20c04efcb0198a3d79171c76983d8a2c1"} Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.499870 4678 scope.go:117] "RemoveContainer" containerID="d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.528887 4678 scope.go:117] "RemoveContainer" containerID="87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.541929 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfgxt"] Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.556199 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rfgxt"] Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.567843 4678 scope.go:117] "RemoveContainer" containerID="44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.584254 4678 scope.go:117] "RemoveContainer" containerID="d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a" Dec 06 10:53:06 crc kubenswrapper[4678]: E1206 10:53:06.584738 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a\": container with ID starting with d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a not found: ID does not exist" containerID="d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.584821 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a"} err="failed to get container status \"d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a\": rpc error: code = NotFound desc = could not find container \"d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a\": container with ID starting with d9b287bfdbe25fb6ac4f197ec53567b3e2fa3b255ca02f5846682bceeab1f51a not found: ID does not exist" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.584852 4678 scope.go:117] "RemoveContainer" containerID="87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5" Dec 06 10:53:06 crc kubenswrapper[4678]: E1206 10:53:06.585722 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5\": container with ID starting with 87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5 not found: ID does not exist" containerID="87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.585782 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5"} err="failed to get container status \"87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5\": rpc error: code = NotFound desc = could not find container \"87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5\": container with ID starting with 87e8c3f27465441f8801d7b0352d7891c8d70550c96a584a84ae5fe1fbcd0cc5 not found: ID does not exist" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.585823 4678 scope.go:117] "RemoveContainer" containerID="44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d" Dec 06 10:53:06 crc kubenswrapper[4678]: E1206 10:53:06.586175 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d\": container with ID starting with 44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d not found: ID does not exist" containerID="44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d" Dec 06 10:53:06 crc kubenswrapper[4678]: I1206 10:53:06.586212 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d"} err="failed to get container status \"44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d\": rpc error: code = NotFound desc = could not find container \"44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d\": container with ID starting with 44562e05281a8b1c6086af8894a16eecff6b7ada12e1e613a1e83f514c5e766d not found: ID does not exist" Dec 06 10:53:07 crc kubenswrapper[4678]: I1206 10:53:07.478253 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 10:53:07 crc kubenswrapper[4678]: I1206 10:53:07.490125 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="106042db-c255-463d-8ea0-60bb39830cb8" path="/var/lib/kubelet/pods/106042db-c255-463d-8ea0-60bb39830cb8/volumes" Dec 06 10:53:07 crc kubenswrapper[4678]: I1206 10:53:07.491403 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" path="/var/lib/kubelet/pods/4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa/volumes" Dec 06 10:53:08 crc kubenswrapper[4678]: I1206 10:53:08.519032 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" event={"ID":"90b306bd-29a2-494d-8399-75d95e9b8597","Type":"ContainerStarted","Data":"ec450e4bb4c89ad8f152ac96a1fef33a7be869bde3fbc18b894ccbc4cfe0e54d"} Dec 06 10:53:08 crc kubenswrapper[4678]: I1206 10:53:08.519730 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" Dec 06 10:53:08 crc kubenswrapper[4678]: I1206 10:53:08.541264 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" podStartSLOduration=3.633547274 podStartE2EDuration="1m14.541239066s" podCreationTimestamp="2025-12-06 10:51:54 +0000 UTC" firstStartedPulling="2025-12-06 10:51:57.058833691 +0000 UTC m=+921.902265130" lastFinishedPulling="2025-12-06 10:53:07.966525483 +0000 UTC m=+992.809956922" observedRunningTime="2025-12-06 10:53:08.537599604 +0000 UTC m=+993.381031043" watchObservedRunningTime="2025-12-06 10:53:08.541239066 +0000 UTC m=+993.384670505" Dec 06 10:53:14 crc kubenswrapper[4678]: I1206 10:53:14.911604 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-wl7ww" Dec 06 10:53:29 crc kubenswrapper[4678]: I1206 10:53:29.993115 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2p8b2"] Dec 06 10:53:29 crc kubenswrapper[4678]: E1206 10:53:29.995508 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="extract-utilities" Dec 06 10:53:29 crc kubenswrapper[4678]: I1206 10:53:29.999824 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="extract-utilities" Dec 06 10:53:29 crc kubenswrapper[4678]: E1206 10:53:29.999963 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="registry-server" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.000034 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="registry-server" Dec 06 10:53:30 crc kubenswrapper[4678]: E1206 10:53:30.000140 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="extract-utilities" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.000219 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="extract-utilities" Dec 06 10:53:30 crc kubenswrapper[4678]: E1206 10:53:30.000327 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="extract-content" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.000403 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="extract-content" Dec 06 10:53:30 crc kubenswrapper[4678]: E1206 10:53:30.000528 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="extract-content" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.000624 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="extract-content" Dec 06 10:53:30 crc kubenswrapper[4678]: E1206 10:53:30.000708 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="registry-server" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.000780 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="registry-server" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.001161 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="106042db-c255-463d-8ea0-60bb39830cb8" containerName="registry-server" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.001200 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d69e5a2-0e60-42b4-a5fe-2ac23947b5aa" containerName="registry-server" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.001995 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.005213 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.005400 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.005560 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.005672 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-xzpcm" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.017938 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2p8b2"] Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.045947 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16ed5b8a-13f8-421a-b754-ff7497efb463-config\") pod \"dnsmasq-dns-675f4bcbfc-2p8b2\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.046192 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwhzh\" (UniqueName: \"kubernetes.io/projected/16ed5b8a-13f8-421a-b754-ff7497efb463-kube-api-access-jwhzh\") pod \"dnsmasq-dns-675f4bcbfc-2p8b2\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.093869 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bqqp"] Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.095400 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.099376 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.124444 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bqqp"] Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.147589 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16ed5b8a-13f8-421a-b754-ff7497efb463-config\") pod \"dnsmasq-dns-675f4bcbfc-2p8b2\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.147970 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfz7g\" (UniqueName: \"kubernetes.io/projected/1de22e17-a9c9-4816-b97c-37d9ff173b40-kube-api-access-vfz7g\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.148102 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-config\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.148231 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwhzh\" (UniqueName: \"kubernetes.io/projected/16ed5b8a-13f8-421a-b754-ff7497efb463-kube-api-access-jwhzh\") pod \"dnsmasq-dns-675f4bcbfc-2p8b2\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.148356 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.149512 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16ed5b8a-13f8-421a-b754-ff7497efb463-config\") pod \"dnsmasq-dns-675f4bcbfc-2p8b2\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.173394 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwhzh\" (UniqueName: \"kubernetes.io/projected/16ed5b8a-13f8-421a-b754-ff7497efb463-kube-api-access-jwhzh\") pod \"dnsmasq-dns-675f4bcbfc-2p8b2\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.249476 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.249644 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfz7g\" (UniqueName: \"kubernetes.io/projected/1de22e17-a9c9-4816-b97c-37d9ff173b40-kube-api-access-vfz7g\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.249682 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-config\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.250361 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.250460 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-config\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.266444 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfz7g\" (UniqueName: \"kubernetes.io/projected/1de22e17-a9c9-4816-b97c-37d9ff173b40-kube-api-access-vfz7g\") pod \"dnsmasq-dns-78dd6ddcc-9bqqp\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.326759 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.418498 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.815319 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2p8b2"] Dec 06 10:53:30 crc kubenswrapper[4678]: W1206 10:53:30.822769 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16ed5b8a_13f8_421a_b754_ff7497efb463.slice/crio-524fa3aa39e618707150c7ea475b9a93f86fcaf5cdf38b6b033a308b204598c6 WatchSource:0}: Error finding container 524fa3aa39e618707150c7ea475b9a93f86fcaf5cdf38b6b033a308b204598c6: Status 404 returned error can't find the container with id 524fa3aa39e618707150c7ea475b9a93f86fcaf5cdf38b6b033a308b204598c6 Dec 06 10:53:30 crc kubenswrapper[4678]: I1206 10:53:30.893276 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bqqp"] Dec 06 10:53:31 crc kubenswrapper[4678]: I1206 10:53:31.729240 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" event={"ID":"16ed5b8a-13f8-421a-b754-ff7497efb463","Type":"ContainerStarted","Data":"524fa3aa39e618707150c7ea475b9a93f86fcaf5cdf38b6b033a308b204598c6"} Dec 06 10:53:31 crc kubenswrapper[4678]: I1206 10:53:31.730734 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" event={"ID":"1de22e17-a9c9-4816-b97c-37d9ff173b40","Type":"ContainerStarted","Data":"b1bcd5e9085f64fa1610adc92cc09787b14869800093ee4b7039037aff04de6c"} Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.073936 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2p8b2"] Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.108284 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fdnth"] Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.113080 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.133517 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fdnth"] Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.211396 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.211570 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf8rq\" (UniqueName: \"kubernetes.io/projected/9ac16ad4-74b1-4968-bd23-10c814ed087d-kube-api-access-rf8rq\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.211602 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-config\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.315124 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf8rq\" (UniqueName: \"kubernetes.io/projected/9ac16ad4-74b1-4968-bd23-10c814ed087d-kube-api-access-rf8rq\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.315201 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-config\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.315231 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.316188 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-dns-svc\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.318581 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-config\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.347723 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf8rq\" (UniqueName: \"kubernetes.io/projected/9ac16ad4-74b1-4968-bd23-10c814ed087d-kube-api-access-rf8rq\") pod \"dnsmasq-dns-666b6646f7-fdnth\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.438888 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bqqp"] Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.450718 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.469362 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f947h"] Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.470817 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.494479 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f947h"] Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.625388 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mfgw\" (UniqueName: \"kubernetes.io/projected/a755af8b-f01a-4434-9de3-33a059a96ba6-kube-api-access-5mfgw\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.625744 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.625789 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-config\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.726856 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mfgw\" (UniqueName: \"kubernetes.io/projected/a755af8b-f01a-4434-9de3-33a059a96ba6-kube-api-access-5mfgw\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.726898 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.726940 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-config\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.727766 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-config\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.727955 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.754672 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mfgw\" (UniqueName: \"kubernetes.io/projected/a755af8b-f01a-4434-9de3-33a059a96ba6-kube-api-access-5mfgw\") pod \"dnsmasq-dns-57d769cc4f-f947h\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:33 crc kubenswrapper[4678]: I1206 10:53:33.897923 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.262399 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fdnth"] Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.305579 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.307288 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.315214 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.315306 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zmxm7" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.315457 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.315614 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.316006 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.316282 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.316299 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.323174 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468201 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468286 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-config-data\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468325 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468375 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468399 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468431 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468461 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468502 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468549 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468578 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.468613 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhk62\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-kube-api-access-rhk62\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.570974 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.571022 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.571061 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.571086 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.571378 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.572174 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.572316 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.573563 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.573648 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhk62\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-kube-api-access-rhk62\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.573872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.573981 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-config-data\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.574094 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.575395 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.577161 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.577845 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.578566 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-config-data\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.579718 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.581193 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.584748 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.585801 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.593965 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f947h"] Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.598069 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhk62\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-kube-api-access-rhk62\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.598183 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.611815 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.648357 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.720405 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.722368 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.722466 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.728195 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.728540 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.728688 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.728803 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.728971 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-p6xng" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.729091 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.730191 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.803892 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.803949 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.803982 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.803999 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.804032 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/995bfee3-8461-4774-aa52-576dc0eacbda-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.804052 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.804073 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.804097 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.804119 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.804138 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksdhv\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-kube-api-access-ksdhv\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.804155 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/995bfee3-8461-4774-aa52-576dc0eacbda-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.819302 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" event={"ID":"a755af8b-f01a-4434-9de3-33a059a96ba6","Type":"ContainerStarted","Data":"68f57a182a8e6ba95ebdf2404c6dfe213aa4229a454e0d5f3f1840d4c161b21b"} Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.832034 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" event={"ID":"9ac16ad4-74b1-4968-bd23-10c814ed087d","Type":"ContainerStarted","Data":"adda73ac535542f65b1d15ddafe2f9fde1f05c91196cdaaf5079600f85f678a2"} Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.905761 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906098 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906138 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksdhv\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-kube-api-access-ksdhv\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906184 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/995bfee3-8461-4774-aa52-576dc0eacbda-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906246 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906280 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906316 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906340 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906383 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/995bfee3-8461-4774-aa52-576dc0eacbda-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.906408 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.911270 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.911876 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.914851 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.916749 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.919252 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.921812 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.925253 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.927800 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.930645 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/995bfee3-8461-4774-aa52-576dc0eacbda-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.931933 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/995bfee3-8461-4774-aa52-576dc0eacbda-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.968542 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:34 crc kubenswrapper[4678]: I1206 10:53:34.968873 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksdhv\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-kube-api-access-ksdhv\") pod \"rabbitmq-cell1-server-0\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:35 crc kubenswrapper[4678]: I1206 10:53:35.066384 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:53:35 crc kubenswrapper[4678]: I1206 10:53:35.320821 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 10:53:35 crc kubenswrapper[4678]: I1206 10:53:35.607629 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 10:53:35 crc kubenswrapper[4678]: I1206 10:53:35.846533 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52efc5a4-585a-4b9e-93f9-f40527e9a6c9","Type":"ContainerStarted","Data":"418dac276e67317edebb84bd03d88223120391184b5c2bad6b3dafd2eea07018"} Dec 06 10:53:35 crc kubenswrapper[4678]: I1206 10:53:35.849593 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"995bfee3-8461-4774-aa52-576dc0eacbda","Type":"ContainerStarted","Data":"80a760531885b8a27aec23e2d66042f4ee95d2b60da078527e4702a040e43caa"} Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.006116 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.007800 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.014102 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6jhp5" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.014405 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.015351 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.015518 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.025505 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.061504 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.126605 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/747aae14-74af-41d1-af19-610b0fa68be1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.126683 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.126706 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/747aae14-74af-41d1-af19-610b0fa68be1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.126756 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747aae14-74af-41d1-af19-610b0fa68be1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.126777 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-config-data-default\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.126972 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t5bp\" (UniqueName: \"kubernetes.io/projected/747aae14-74af-41d1-af19-610b0fa68be1-kube-api-access-7t5bp\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.127060 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-kolla-config\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.127084 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.230593 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.233580 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.233843 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/747aae14-74af-41d1-af19-610b0fa68be1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.233969 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747aae14-74af-41d1-af19-610b0fa68be1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.234074 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-config-data-default\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.234182 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t5bp\" (UniqueName: \"kubernetes.io/projected/747aae14-74af-41d1-af19-610b0fa68be1-kube-api-access-7t5bp\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.234527 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-kolla-config\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.234636 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.234738 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/747aae14-74af-41d1-af19-610b0fa68be1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.236723 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/747aae14-74af-41d1-af19-610b0fa68be1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.237393 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-kolla-config\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.237813 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-config-data-default\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.274359 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/747aae14-74af-41d1-af19-610b0fa68be1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.275153 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/747aae14-74af-41d1-af19-610b0fa68be1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.276093 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/747aae14-74af-41d1-af19-610b0fa68be1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.357409 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t5bp\" (UniqueName: \"kubernetes.io/projected/747aae14-74af-41d1-af19-610b0fa68be1-kube-api-access-7t5bp\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.434980 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"747aae14-74af-41d1-af19-610b0fa68be1\") " pod="openstack/openstack-galera-0" Dec 06 10:53:36 crc kubenswrapper[4678]: I1206 10:53:36.662895 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.383166 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.385410 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.388379 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.388687 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.388952 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-bkzwc" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.389093 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.412375 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.457903 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1cb14647-bb80-422d-8f59-985a70bf14c3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.457948 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.457987 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.458006 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.458030 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncbzd\" (UniqueName: \"kubernetes.io/projected/1cb14647-bb80-422d-8f59-985a70bf14c3-kube-api-access-ncbzd\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.458066 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.458091 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb14647-bb80-422d-8f59-985a70bf14c3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.458110 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb14647-bb80-422d-8f59-985a70bf14c3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561296 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561364 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb14647-bb80-422d-8f59-985a70bf14c3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561386 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb14647-bb80-422d-8f59-985a70bf14c3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561431 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1cb14647-bb80-422d-8f59-985a70bf14c3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561450 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561479 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561514 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.561531 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncbzd\" (UniqueName: \"kubernetes.io/projected/1cb14647-bb80-422d-8f59-985a70bf14c3-kube-api-access-ncbzd\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.562855 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.564036 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.566819 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.567968 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1cb14647-bb80-422d-8f59-985a70bf14c3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.572284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1cb14647-bb80-422d-8f59-985a70bf14c3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.590280 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cb14647-bb80-422d-8f59-985a70bf14c3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.594810 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1cb14647-bb80-422d-8f59-985a70bf14c3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.603005 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncbzd\" (UniqueName: \"kubernetes.io/projected/1cb14647-bb80-422d-8f59-985a70bf14c3-kube-api-access-ncbzd\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.639237 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1cb14647-bb80-422d-8f59-985a70bf14c3\") " pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.731931 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.813639 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.814866 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.828826 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-l2zv4" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.829038 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.841451 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.859080 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.870287 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt2zz\" (UniqueName: \"kubernetes.io/projected/a0c958d8-87e6-4b93-be23-58cb84d32dd4-kube-api-access-bt2zz\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.870362 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c958d8-87e6-4b93-be23-58cb84d32dd4-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.870415 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0c958d8-87e6-4b93-be23-58cb84d32dd4-kolla-config\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.870429 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c958d8-87e6-4b93-be23-58cb84d32dd4-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.870457 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0c958d8-87e6-4b93-be23-58cb84d32dd4-config-data\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.961253 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.973122 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0c958d8-87e6-4b93-be23-58cb84d32dd4-config-data\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.973233 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt2zz\" (UniqueName: \"kubernetes.io/projected/a0c958d8-87e6-4b93-be23-58cb84d32dd4-kube-api-access-bt2zz\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.973285 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c958d8-87e6-4b93-be23-58cb84d32dd4-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.973340 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0c958d8-87e6-4b93-be23-58cb84d32dd4-kolla-config\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.973361 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c958d8-87e6-4b93-be23-58cb84d32dd4-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.987122 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a0c958d8-87e6-4b93-be23-58cb84d32dd4-config-data\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.987675 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a0c958d8-87e6-4b93-be23-58cb84d32dd4-kolla-config\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.989665 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0c958d8-87e6-4b93-be23-58cb84d32dd4-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:37 crc kubenswrapper[4678]: I1206 10:53:37.990089 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c958d8-87e6-4b93-be23-58cb84d32dd4-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:38 crc kubenswrapper[4678]: I1206 10:53:38.008070 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt2zz\" (UniqueName: \"kubernetes.io/projected/a0c958d8-87e6-4b93-be23-58cb84d32dd4-kube-api-access-bt2zz\") pod \"memcached-0\" (UID: \"a0c958d8-87e6-4b93-be23-58cb84d32dd4\") " pod="openstack/memcached-0" Dec 06 10:53:38 crc kubenswrapper[4678]: I1206 10:53:38.152205 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 06 10:53:38 crc kubenswrapper[4678]: I1206 10:53:38.467053 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 06 10:53:38 crc kubenswrapper[4678]: I1206 10:53:38.785691 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 06 10:53:38 crc kubenswrapper[4678]: I1206 10:53:38.968642 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"747aae14-74af-41d1-af19-610b0fa68be1","Type":"ContainerStarted","Data":"bce3c4c1bdcabe086bd0462675756e1244b9288a6fc89cb4fb7a58304436d598"} Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.119839 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.121162 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.123347 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nmq8v" Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.149115 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.322636 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6bln\" (UniqueName: \"kubernetes.io/projected/6e0fe084-357f-4b1e-bf7a-bf0413e830b1-kube-api-access-l6bln\") pod \"kube-state-metrics-0\" (UID: \"6e0fe084-357f-4b1e-bf7a-bf0413e830b1\") " pod="openstack/kube-state-metrics-0" Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.426468 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6bln\" (UniqueName: \"kubernetes.io/projected/6e0fe084-357f-4b1e-bf7a-bf0413e830b1-kube-api-access-l6bln\") pod \"kube-state-metrics-0\" (UID: \"6e0fe084-357f-4b1e-bf7a-bf0413e830b1\") " pod="openstack/kube-state-metrics-0" Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.469783 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6bln\" (UniqueName: \"kubernetes.io/projected/6e0fe084-357f-4b1e-bf7a-bf0413e830b1-kube-api-access-l6bln\") pod \"kube-state-metrics-0\" (UID: \"6e0fe084-357f-4b1e-bf7a-bf0413e830b1\") " pod="openstack/kube-state-metrics-0" Dec 06 10:53:40 crc kubenswrapper[4678]: I1206 10:53:40.762726 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.865783 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kdb5q"] Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.867421 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.871979 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.871979 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.872346 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-c448k" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.876038 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-zcg6x"] Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.878023 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.888211 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kdb5q"] Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909294 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2658168-1355-4595-bf23-570a68c90da1-ovn-controller-tls-certs\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909382 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82klq\" (UniqueName: \"kubernetes.io/projected/0155b73c-73e9-4b08-830d-f78f260d6ae5-kube-api-access-82klq\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909422 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-run\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909478 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-log\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909515 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-run-ovn\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909547 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-run\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909584 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2658168-1355-4595-bf23-570a68c90da1-combined-ca-bundle\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909644 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0155b73c-73e9-4b08-830d-f78f260d6ae5-scripts\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909698 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fns8j\" (UniqueName: \"kubernetes.io/projected/c2658168-1355-4595-bf23-570a68c90da1-kube-api-access-fns8j\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909729 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-etc-ovs\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909802 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-lib\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909830 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2658168-1355-4595-bf23-570a68c90da1-scripts\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.909879 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-log-ovn\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:44 crc kubenswrapper[4678]: I1206 10:53:44.918352 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zcg6x"] Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010675 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2658168-1355-4595-bf23-570a68c90da1-combined-ca-bundle\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010727 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0155b73c-73e9-4b08-830d-f78f260d6ae5-scripts\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010769 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fns8j\" (UniqueName: \"kubernetes.io/projected/c2658168-1355-4595-bf23-570a68c90da1-kube-api-access-fns8j\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010796 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-etc-ovs\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010819 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-lib\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010840 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2658168-1355-4595-bf23-570a68c90da1-scripts\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010858 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-log-ovn\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010882 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2658168-1355-4595-bf23-570a68c90da1-ovn-controller-tls-certs\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010914 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82klq\" (UniqueName: \"kubernetes.io/projected/0155b73c-73e9-4b08-830d-f78f260d6ae5-kube-api-access-82klq\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010931 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-run\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010960 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-log\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010976 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-run-ovn\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.010996 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-run\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.011574 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-run\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.011628 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-run\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.011701 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-etc-ovs\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.011780 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-run-ovn\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.012364 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-log\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.013833 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0155b73c-73e9-4b08-830d-f78f260d6ae5-scripts\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.013951 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c2658168-1355-4595-bf23-570a68c90da1-var-log-ovn\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.014719 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2658168-1355-4595-bf23-570a68c90da1-scripts\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.015650 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/0155b73c-73e9-4b08-830d-f78f260d6ae5-var-lib\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.016587 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2658168-1355-4595-bf23-570a68c90da1-combined-ca-bundle\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.017008 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2658168-1355-4595-bf23-570a68c90da1-ovn-controller-tls-certs\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.031478 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fns8j\" (UniqueName: \"kubernetes.io/projected/c2658168-1355-4595-bf23-570a68c90da1-kube-api-access-fns8j\") pod \"ovn-controller-kdb5q\" (UID: \"c2658168-1355-4595-bf23-570a68c90da1\") " pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.041442 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82klq\" (UniqueName: \"kubernetes.io/projected/0155b73c-73e9-4b08-830d-f78f260d6ae5-kube-api-access-82klq\") pod \"ovn-controller-ovs-zcg6x\" (UID: \"0155b73c-73e9-4b08-830d-f78f260d6ae5\") " pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.195730 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kdb5q" Dec 06 10:53:45 crc kubenswrapper[4678]: I1206 10:53:45.215721 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.114307 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.116270 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.118685 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.118860 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.118860 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.119440 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-625rd" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.120287 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.136983 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266422 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266467 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-config\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266527 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266555 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266665 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266799 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rngg\" (UniqueName: \"kubernetes.io/projected/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-kube-api-access-8rngg\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266846 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.266875 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.305965 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.307397 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.311267 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-4c5hv" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.312674 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.313118 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.320969 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.334247 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368608 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rngg\" (UniqueName: \"kubernetes.io/projected/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-kube-api-access-8rngg\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368692 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368718 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368743 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368760 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-config\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368798 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368820 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.368847 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.369217 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.370136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.370247 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-config\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.370502 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.379921 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.380031 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.387042 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.394017 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.394170 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rngg\" (UniqueName: \"kubernetes.io/projected/12052ef6-7a13-4ad5-9484-bbad2cabdf5a-kube-api-access-8rngg\") pod \"ovsdbserver-nb-0\" (UID: \"12052ef6-7a13-4ad5-9484-bbad2cabdf5a\") " pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.434045 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470226 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6c33f7-7b74-437b-a9fd-3790af864af7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470275 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f6c33f7-7b74-437b-a9fd-3790af864af7-config\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470304 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brmlw\" (UniqueName: \"kubernetes.io/projected/3f6c33f7-7b74-437b-a9fd-3790af864af7-kube-api-access-brmlw\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470420 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470475 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470515 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470544 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3f6c33f7-7b74-437b-a9fd-3790af864af7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.470712 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.572683 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.572790 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.572858 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3f6c33f7-7b74-437b-a9fd-3790af864af7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.572890 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.572907 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.573012 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6c33f7-7b74-437b-a9fd-3790af864af7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.573120 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f6c33f7-7b74-437b-a9fd-3790af864af7-config\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.574125 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f6c33f7-7b74-437b-a9fd-3790af864af7-config\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.573427 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3f6c33f7-7b74-437b-a9fd-3790af864af7-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.574670 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brmlw\" (UniqueName: \"kubernetes.io/projected/3f6c33f7-7b74-437b-a9fd-3790af864af7-kube-api-access-brmlw\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.574711 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.575192 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f6c33f7-7b74-437b-a9fd-3790af864af7-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.580113 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.581276 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.586762 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f6c33f7-7b74-437b-a9fd-3790af864af7-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.592176 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brmlw\" (UniqueName: \"kubernetes.io/projected/3f6c33f7-7b74-437b-a9fd-3790af864af7-kube-api-access-brmlw\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.596845 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-sb-0\" (UID: \"3f6c33f7-7b74-437b-a9fd-3790af864af7\") " pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:50 crc kubenswrapper[4678]: I1206 10:53:47.624787 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 06 10:53:51 crc kubenswrapper[4678]: I1206 10:53:51.128766 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1cb14647-bb80-422d-8f59-985a70bf14c3","Type":"ContainerStarted","Data":"675d1374b9b251887726574c22ce6118a87039f72420fad8be8de932dc239206"} Dec 06 10:53:51 crc kubenswrapper[4678]: I1206 10:53:51.132035 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a0c958d8-87e6-4b93-be23-58cb84d32dd4","Type":"ContainerStarted","Data":"ecec398ab61039671f063b36298d4b108d079680cd2a2d6349f1418f99cc8fad"} Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.397154 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.398447 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5mfgw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-f947h_openstack(a755af8b-f01a-4434-9de3-33a059a96ba6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.399854 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.401790 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.401950 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rf8rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-fdnth_openstack(9ac16ad4-74b1-4968-bd23-10c814ed087d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.403384 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.416239 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.416403 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vfz7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-9bqqp_openstack(1de22e17-a9c9-4816-b97c-37d9ff173b40): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.417543 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" podUID="1de22e17-a9c9-4816-b97c-37d9ff173b40" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.450914 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.451085 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jwhzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-2p8b2_openstack(16ed5b8a-13f8-421a-b754-ff7497efb463): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:53:59 crc kubenswrapper[4678]: E1206 10:53:59.452466 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" podUID="16ed5b8a-13f8-421a-b754-ff7497efb463" Dec 06 10:54:00 crc kubenswrapper[4678]: E1206 10:54:00.201201 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" Dec 06 10:54:00 crc kubenswrapper[4678]: E1206 10:54:00.201134 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" Dec 06 10:54:01 crc kubenswrapper[4678]: E1206 10:54:01.008107 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 06 10:54:01 crc kubenswrapper[4678]: E1206 10:54:01.008978 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rhk62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(52efc5a4-585a-4b9e-93f9-f40527e9a6c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:54:01 crc kubenswrapper[4678]: E1206 10:54:01.010231 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" Dec 06 10:54:01 crc kubenswrapper[4678]: E1206 10:54:01.207809 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" Dec 06 10:54:04 crc kubenswrapper[4678]: I1206 10:54:04.903691 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:54:04 crc kubenswrapper[4678]: I1206 10:54:04.910528 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.009341 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwhzh\" (UniqueName: \"kubernetes.io/projected/16ed5b8a-13f8-421a-b754-ff7497efb463-kube-api-access-jwhzh\") pod \"16ed5b8a-13f8-421a-b754-ff7497efb463\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.009392 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16ed5b8a-13f8-421a-b754-ff7497efb463-config\") pod \"16ed5b8a-13f8-421a-b754-ff7497efb463\" (UID: \"16ed5b8a-13f8-421a-b754-ff7497efb463\") " Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.009434 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfz7g\" (UniqueName: \"kubernetes.io/projected/1de22e17-a9c9-4816-b97c-37d9ff173b40-kube-api-access-vfz7g\") pod \"1de22e17-a9c9-4816-b97c-37d9ff173b40\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.009536 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-dns-svc\") pod \"1de22e17-a9c9-4816-b97c-37d9ff173b40\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.010073 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16ed5b8a-13f8-421a-b754-ff7497efb463-config" (OuterVolumeSpecName: "config") pod "16ed5b8a-13f8-421a-b754-ff7497efb463" (UID: "16ed5b8a-13f8-421a-b754-ff7497efb463"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.010385 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-config\") pod \"1de22e17-a9c9-4816-b97c-37d9ff173b40\" (UID: \"1de22e17-a9c9-4816-b97c-37d9ff173b40\") " Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.010781 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1de22e17-a9c9-4816-b97c-37d9ff173b40" (UID: "1de22e17-a9c9-4816-b97c-37d9ff173b40"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.010903 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16ed5b8a-13f8-421a-b754-ff7497efb463-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.011526 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-config" (OuterVolumeSpecName: "config") pod "1de22e17-a9c9-4816-b97c-37d9ff173b40" (UID: "1de22e17-a9c9-4816-b97c-37d9ff173b40"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.016437 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1de22e17-a9c9-4816-b97c-37d9ff173b40-kube-api-access-vfz7g" (OuterVolumeSpecName: "kube-api-access-vfz7g") pod "1de22e17-a9c9-4816-b97c-37d9ff173b40" (UID: "1de22e17-a9c9-4816-b97c-37d9ff173b40"). InnerVolumeSpecName "kube-api-access-vfz7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.030966 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16ed5b8a-13f8-421a-b754-ff7497efb463-kube-api-access-jwhzh" (OuterVolumeSpecName: "kube-api-access-jwhzh") pod "16ed5b8a-13f8-421a-b754-ff7497efb463" (UID: "16ed5b8a-13f8-421a-b754-ff7497efb463"). InnerVolumeSpecName "kube-api-access-jwhzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.112835 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.112868 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1de22e17-a9c9-4816-b97c-37d9ff173b40-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.112879 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwhzh\" (UniqueName: \"kubernetes.io/projected/16ed5b8a-13f8-421a-b754-ff7497efb463-kube-api-access-jwhzh\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.112891 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfz7g\" (UniqueName: \"kubernetes.io/projected/1de22e17-a9c9-4816-b97c-37d9ff173b40-kube-api-access-vfz7g\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.252283 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.252272 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-9bqqp" event={"ID":"1de22e17-a9c9-4816-b97c-37d9ff173b40","Type":"ContainerDied","Data":"b1bcd5e9085f64fa1610adc92cc09787b14869800093ee4b7039037aff04de6c"} Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.256520 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" event={"ID":"16ed5b8a-13f8-421a-b754-ff7497efb463","Type":"ContainerDied","Data":"524fa3aa39e618707150c7ea475b9a93f86fcaf5cdf38b6b033a308b204598c6"} Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.256590 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-2p8b2" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.355584 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bqqp"] Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.368471 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-9bqqp"] Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.382746 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2p8b2"] Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.391405 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-2p8b2"] Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.498481 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16ed5b8a-13f8-421a-b754-ff7497efb463" path="/var/lib/kubelet/pods/16ed5b8a-13f8-421a-b754-ff7497efb463/volumes" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.498893 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1de22e17-a9c9-4816-b97c-37d9ff173b40" path="/var/lib/kubelet/pods/1de22e17-a9c9-4816-b97c-37d9ff173b40/volumes" Dec 06 10:54:05 crc kubenswrapper[4678]: I1206 10:54:05.499186 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 06 10:54:06 crc kubenswrapper[4678]: I1206 10:54:06.265273 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3f6c33f7-7b74-437b-a9fd-3790af864af7","Type":"ContainerStarted","Data":"630eb98354b498277e618afea41c67f89ae9c01f0773a5682097068145ab5442"} Dec 06 10:54:06 crc kubenswrapper[4678]: I1206 10:54:06.466436 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:54:06 crc kubenswrapper[4678]: W1206 10:54:06.476476 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e0fe084_357f_4b1e_bf7a_bf0413e830b1.slice/crio-3c99e6479bd2aab8cd7606325e03c75a936d6d9fbbadb37b8a942c909b747445 WatchSource:0}: Error finding container 3c99e6479bd2aab8cd7606325e03c75a936d6d9fbbadb37b8a942c909b747445: Status 404 returned error can't find the container with id 3c99e6479bd2aab8cd7606325e03c75a936d6d9fbbadb37b8a942c909b747445 Dec 06 10:54:06 crc kubenswrapper[4678]: I1206 10:54:06.562545 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 06 10:54:06 crc kubenswrapper[4678]: W1206 10:54:06.568778 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12052ef6_7a13_4ad5_9484_bbad2cabdf5a.slice/crio-3179e432e23b91ec4162b7860ddf0f0caa1f597d72b20709f6ac01c55ff1f33e WatchSource:0}: Error finding container 3179e432e23b91ec4162b7860ddf0f0caa1f597d72b20709f6ac01c55ff1f33e: Status 404 returned error can't find the container with id 3179e432e23b91ec4162b7860ddf0f0caa1f597d72b20709f6ac01c55ff1f33e Dec 06 10:54:06 crc kubenswrapper[4678]: I1206 10:54:06.574644 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kdb5q"] Dec 06 10:54:06 crc kubenswrapper[4678]: I1206 10:54:06.647286 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zcg6x"] Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.285779 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"995bfee3-8461-4774-aa52-576dc0eacbda","Type":"ContainerStarted","Data":"a67896673bf325c989b177f147cde987c7fc2280ee8ac8bf78a9c554bbb3c2b7"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.289970 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kdb5q" event={"ID":"c2658168-1355-4595-bf23-570a68c90da1","Type":"ContainerStarted","Data":"cd860e798a59b617510877a4e90ab3e40b27c70e4febb5abdbdb040ebe9464ab"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.304120 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a0c958d8-87e6-4b93-be23-58cb84d32dd4","Type":"ContainerStarted","Data":"9ad52acfdf99873c81d71e486791103042c12ef922eb32f32d844bac329b6d4b"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.304325 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.305505 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"12052ef6-7a13-4ad5-9484-bbad2cabdf5a","Type":"ContainerStarted","Data":"3179e432e23b91ec4162b7860ddf0f0caa1f597d72b20709f6ac01c55ff1f33e"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.306931 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6e0fe084-357f-4b1e-bf7a-bf0413e830b1","Type":"ContainerStarted","Data":"3c99e6479bd2aab8cd7606325e03c75a936d6d9fbbadb37b8a942c909b747445"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.320414 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"747aae14-74af-41d1-af19-610b0fa68be1","Type":"ContainerStarted","Data":"6c2b8fba08f4e41db8234b26d6391e8dcbef361e30b073f5131987f156c8cba7"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.324041 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zcg6x" event={"ID":"0155b73c-73e9-4b08-830d-f78f260d6ae5","Type":"ContainerStarted","Data":"b5f21e3c35ca5c56f559baa5d2faf1b60712a27fd64de79c925007f265f90778"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.327915 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1cb14647-bb80-422d-8f59-985a70bf14c3","Type":"ContainerStarted","Data":"df640bc1782ef2979032e010418a4ebeabe6116707a2b6da9ac6bc4cf83708cb"} Dec 06 10:54:07 crc kubenswrapper[4678]: I1206 10:54:07.341134 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=15.420485891 podStartE2EDuration="30.341115082s" podCreationTimestamp="2025-12-06 10:53:37 +0000 UTC" firstStartedPulling="2025-12-06 10:53:51.095634749 +0000 UTC m=+1035.939066188" lastFinishedPulling="2025-12-06 10:54:06.01626394 +0000 UTC m=+1050.859695379" observedRunningTime="2025-12-06 10:54:07.33418704 +0000 UTC m=+1052.177618479" watchObservedRunningTime="2025-12-06 10:54:07.341115082 +0000 UTC m=+1052.184546521" Dec 06 10:54:13 crc kubenswrapper[4678]: I1206 10:54:13.157319 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 06 10:54:14 crc kubenswrapper[4678]: I1206 10:54:14.388959 4678 generic.go:334] "Generic (PLEG): container finished" podID="1cb14647-bb80-422d-8f59-985a70bf14c3" containerID="df640bc1782ef2979032e010418a4ebeabe6116707a2b6da9ac6bc4cf83708cb" exitCode=0 Dec 06 10:54:14 crc kubenswrapper[4678]: I1206 10:54:14.389070 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1cb14647-bb80-422d-8f59-985a70bf14c3","Type":"ContainerDied","Data":"df640bc1782ef2979032e010418a4ebeabe6116707a2b6da9ac6bc4cf83708cb"} Dec 06 10:54:15 crc kubenswrapper[4678]: I1206 10:54:15.451454 4678 generic.go:334] "Generic (PLEG): container finished" podID="747aae14-74af-41d1-af19-610b0fa68be1" containerID="6c2b8fba08f4e41db8234b26d6391e8dcbef361e30b073f5131987f156c8cba7" exitCode=0 Dec 06 10:54:15 crc kubenswrapper[4678]: I1206 10:54:15.451538 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"747aae14-74af-41d1-af19-610b0fa68be1","Type":"ContainerDied","Data":"6c2b8fba08f4e41db8234b26d6391e8dcbef361e30b073f5131987f156c8cba7"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.469665 4678 generic.go:334] "Generic (PLEG): container finished" podID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerID="0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce" exitCode=0 Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.470473 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" event={"ID":"a755af8b-f01a-4434-9de3-33a059a96ba6","Type":"ContainerDied","Data":"0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.484174 4678 generic.go:334] "Generic (PLEG): container finished" podID="0155b73c-73e9-4b08-830d-f78f260d6ae5" containerID="1e59181edcd6b87e3a4e4a2b651ce7301e2f99671e7fcd42ff2ee6847a673488" exitCode=0 Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.502437 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"12052ef6-7a13-4ad5-9484-bbad2cabdf5a","Type":"ContainerStarted","Data":"0bed9d903cf806d7cf2b538a04887a28d0c3db0ed8dbb5cc5c90033a37898de0"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.502614 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3f6c33f7-7b74-437b-a9fd-3790af864af7","Type":"ContainerStarted","Data":"88bcd15f778921ba9223066ed569fd20a18312d28a3d7ef8eb20793335386dc4"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.502729 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"747aae14-74af-41d1-af19-610b0fa68be1","Type":"ContainerStarted","Data":"a69436c6317ff6ef4b68ee84babb54f6c739d818be9bc56038e2e9b906e21bbe"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.502832 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zcg6x" event={"ID":"0155b73c-73e9-4b08-830d-f78f260d6ae5","Type":"ContainerDied","Data":"1e59181edcd6b87e3a4e4a2b651ce7301e2f99671e7fcd42ff2ee6847a673488"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.502922 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1cb14647-bb80-422d-8f59-985a70bf14c3","Type":"ContainerStarted","Data":"41d1a0932e16be45e65034ad6f72ba3be985c70546f595f6d49323cb779e4c0c"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.523664 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kdb5q" event={"ID":"c2658168-1355-4595-bf23-570a68c90da1","Type":"ContainerStarted","Data":"7183a2cd8193e84bc495959b751b9d7edd4379b90093d75cafd6f006814c46ec"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.527187 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-kdb5q" Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.546426 4678 generic.go:334] "Generic (PLEG): container finished" podID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerID="0b3242f24bd29cf8c95a52f2852b84550f148909fe5ac46f4627d3cd7805dbf6" exitCode=0 Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.546521 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" event={"ID":"9ac16ad4-74b1-4968-bd23-10c814ed087d","Type":"ContainerDied","Data":"0b3242f24bd29cf8c95a52f2852b84550f148909fe5ac46f4627d3cd7805dbf6"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.550936 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6e0fe084-357f-4b1e-bf7a-bf0413e830b1","Type":"ContainerStarted","Data":"934f5ddaf048b192feb57a1ea23b3ae827e983513df5661e44fa5b2a02a46acf"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.551159 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.552255 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=15.622121673 podStartE2EDuration="43.552242306s" podCreationTimestamp="2025-12-06 10:53:34 +0000 UTC" firstStartedPulling="2025-12-06 10:53:38.014039587 +0000 UTC m=+1022.857471016" lastFinishedPulling="2025-12-06 10:54:05.94416021 +0000 UTC m=+1050.787591649" observedRunningTime="2025-12-06 10:54:17.542563709 +0000 UTC m=+1062.385995168" watchObservedRunningTime="2025-12-06 10:54:17.552242306 +0000 UTC m=+1062.395673745" Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.559944 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52efc5a4-585a-4b9e-93f9-f40527e9a6c9","Type":"ContainerStarted","Data":"f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b"} Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.580869 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=26.667303399 podStartE2EDuration="41.580847723s" podCreationTimestamp="2025-12-06 10:53:36 +0000 UTC" firstStartedPulling="2025-12-06 10:53:51.104259443 +0000 UTC m=+1035.947690882" lastFinishedPulling="2025-12-06 10:54:06.017803767 +0000 UTC m=+1050.861235206" observedRunningTime="2025-12-06 10:54:17.570584257 +0000 UTC m=+1062.414015706" watchObservedRunningTime="2025-12-06 10:54:17.580847723 +0000 UTC m=+1062.424279152" Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.633594 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=27.392481985 podStartE2EDuration="37.633469005s" podCreationTimestamp="2025-12-06 10:53:40 +0000 UTC" firstStartedPulling="2025-12-06 10:54:06.482181958 +0000 UTC m=+1051.325613397" lastFinishedPulling="2025-12-06 10:54:16.723168978 +0000 UTC m=+1061.566600417" observedRunningTime="2025-12-06 10:54:17.616032901 +0000 UTC m=+1062.459464340" watchObservedRunningTime="2025-12-06 10:54:17.633469005 +0000 UTC m=+1062.476900444" Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.700641 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-kdb5q" podStartSLOduration=25.028727511 podStartE2EDuration="33.700617033s" podCreationTimestamp="2025-12-06 10:53:44 +0000 UTC" firstStartedPulling="2025-12-06 10:54:06.583874885 +0000 UTC m=+1051.427306324" lastFinishedPulling="2025-12-06 10:54:15.255764377 +0000 UTC m=+1060.099195846" observedRunningTime="2025-12-06 10:54:17.694792495 +0000 UTC m=+1062.538223954" watchObservedRunningTime="2025-12-06 10:54:17.700617033 +0000 UTC m=+1062.544048472" Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.732757 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 06 10:54:17 crc kubenswrapper[4678]: I1206 10:54:17.732885 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.573999 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zcg6x" event={"ID":"0155b73c-73e9-4b08-830d-f78f260d6ae5","Type":"ContainerStarted","Data":"24b16415b9c2d9bbd2427001238578ca7d2897cf0d08a61346c418b16da8126a"} Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.574040 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zcg6x" event={"ID":"0155b73c-73e9-4b08-830d-f78f260d6ae5","Type":"ContainerStarted","Data":"f129d05e1f596e062baed668146a89b683f9c0bdb4b132d97b21c1d6b1c0f03a"} Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.574386 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.574504 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.586138 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" event={"ID":"a755af8b-f01a-4434-9de3-33a059a96ba6","Type":"ContainerStarted","Data":"98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece"} Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.586735 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.592691 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" event={"ID":"9ac16ad4-74b1-4968-bd23-10c814ed087d","Type":"ContainerStarted","Data":"2c94b4c0a6c2dbabaf6944ff038082789e635ac8c153e525d547a63b452a54bd"} Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.593199 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.596067 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-zcg6x" podStartSLOduration=26.154831291 podStartE2EDuration="34.596047384s" podCreationTimestamp="2025-12-06 10:53:44 +0000 UTC" firstStartedPulling="2025-12-06 10:54:06.658988147 +0000 UTC m=+1051.502419586" lastFinishedPulling="2025-12-06 10:54:15.10020422 +0000 UTC m=+1059.943635679" observedRunningTime="2025-12-06 10:54:18.594214969 +0000 UTC m=+1063.437646408" watchObservedRunningTime="2025-12-06 10:54:18.596047384 +0000 UTC m=+1063.439478823" Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.614664 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" podStartSLOduration=3.46355626 podStartE2EDuration="45.614617653s" podCreationTimestamp="2025-12-06 10:53:33 +0000 UTC" firstStartedPulling="2025-12-06 10:53:34.601751873 +0000 UTC m=+1019.445183312" lastFinishedPulling="2025-12-06 10:54:16.752813276 +0000 UTC m=+1061.596244705" observedRunningTime="2025-12-06 10:54:18.611756946 +0000 UTC m=+1063.455188385" watchObservedRunningTime="2025-12-06 10:54:18.614617653 +0000 UTC m=+1063.458049082" Dec 06 10:54:18 crc kubenswrapper[4678]: I1206 10:54:18.630963 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" podStartSLOduration=3.176522614 podStartE2EDuration="45.630931624s" podCreationTimestamp="2025-12-06 10:53:33 +0000 UTC" firstStartedPulling="2025-12-06 10:53:34.292063802 +0000 UTC m=+1019.135495241" lastFinishedPulling="2025-12-06 10:54:16.746472822 +0000 UTC m=+1061.589904251" observedRunningTime="2025-12-06 10:54:18.62820399 +0000 UTC m=+1063.471635429" watchObservedRunningTime="2025-12-06 10:54:18.630931624 +0000 UTC m=+1063.474363063" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.238397 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f947h"] Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.293186 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-9khdg"] Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.294812 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.309877 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-9khdg"] Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.335867 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-config\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.336040 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9b7zw\" (UniqueName: \"kubernetes.io/projected/04529fe4-5999-4649-8dff-674193cd5f1e-kube-api-access-9b7zw\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.336077 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.437404 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9b7zw\" (UniqueName: \"kubernetes.io/projected/04529fe4-5999-4649-8dff-674193cd5f1e-kube-api-access-9b7zw\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.437456 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.437554 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-config\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.438326 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-config\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.438833 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.475675 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9b7zw\" (UniqueName: \"kubernetes.io/projected/04529fe4-5999-4649-8dff-674193cd5f1e-kube-api-access-9b7zw\") pod \"dnsmasq-dns-7cb5889db5-9khdg\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.607074 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerName="dnsmasq-dns" containerID="cri-o://98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece" gracePeriod=10 Dec 06 10:54:20 crc kubenswrapper[4678]: I1206 10:54:20.615313 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.245092 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.276830 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-9khdg"] Dec 06 10:54:21 crc kubenswrapper[4678]: W1206 10:54:21.292767 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04529fe4_5999_4649_8dff_674193cd5f1e.slice/crio-934b10aa291b7465e3c379fc8fc70d05509fc499db17f30873700c4fda8bbebb WatchSource:0}: Error finding container 934b10aa291b7465e3c379fc8fc70d05509fc499db17f30873700c4fda8bbebb: Status 404 returned error can't find the container with id 934b10aa291b7465e3c379fc8fc70d05509fc499db17f30873700c4fda8bbebb Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.351570 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-config\") pod \"a755af8b-f01a-4434-9de3-33a059a96ba6\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.351625 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-dns-svc\") pod \"a755af8b-f01a-4434-9de3-33a059a96ba6\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.351681 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mfgw\" (UniqueName: \"kubernetes.io/projected/a755af8b-f01a-4434-9de3-33a059a96ba6-kube-api-access-5mfgw\") pod \"a755af8b-f01a-4434-9de3-33a059a96ba6\" (UID: \"a755af8b-f01a-4434-9de3-33a059a96ba6\") " Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.353134 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 06 10:54:21 crc kubenswrapper[4678]: E1206 10:54:21.353711 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerName="dnsmasq-dns" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.353728 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerName="dnsmasq-dns" Dec 06 10:54:21 crc kubenswrapper[4678]: E1206 10:54:21.353762 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerName="init" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.353769 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerName="init" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.353963 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerName="dnsmasq-dns" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.364741 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.368620 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.368681 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.368989 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.370775 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-mvhj5" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.373886 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a755af8b-f01a-4434-9de3-33a059a96ba6-kube-api-access-5mfgw" (OuterVolumeSpecName: "kube-api-access-5mfgw") pod "a755af8b-f01a-4434-9de3-33a059a96ba6" (UID: "a755af8b-f01a-4434-9de3-33a059a96ba6"). InnerVolumeSpecName "kube-api-access-5mfgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.376417 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.453891 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq5q9\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-kube-api-access-cq5q9\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.454086 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.454157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a7f4ca97-1480-4327-8025-9b18a8a7073b-lock\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.454252 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a7f4ca97-1480-4327-8025-9b18a8a7073b-cache\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.454283 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.454420 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mfgw\" (UniqueName: \"kubernetes.io/projected/a755af8b-f01a-4434-9de3-33a059a96ba6-kube-api-access-5mfgw\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.468817 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-config" (OuterVolumeSpecName: "config") pod "a755af8b-f01a-4434-9de3-33a059a96ba6" (UID: "a755af8b-f01a-4434-9de3-33a059a96ba6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.471601 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a755af8b-f01a-4434-9de3-33a059a96ba6" (UID: "a755af8b-f01a-4434-9de3-33a059a96ba6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.555561 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq5q9\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-kube-api-access-cq5q9\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.555928 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.555966 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a7f4ca97-1480-4327-8025-9b18a8a7073b-lock\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.556018 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a7f4ca97-1480-4327-8025-9b18a8a7073b-cache\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.556049 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.556111 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.556150 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a755af8b-f01a-4434-9de3-33a059a96ba6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.556477 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.556866 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a7f4ca97-1480-4327-8025-9b18a8a7073b-cache\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.557349 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/a7f4ca97-1480-4327-8025-9b18a8a7073b-lock\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: E1206 10:54:21.557440 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 10:54:21 crc kubenswrapper[4678]: E1206 10:54:21.557461 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 10:54:21 crc kubenswrapper[4678]: E1206 10:54:21.557607 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift podName:a7f4ca97-1480-4327-8025-9b18a8a7073b nodeName:}" failed. No retries permitted until 2025-12-06 10:54:22.057588814 +0000 UTC m=+1066.901020253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift") pod "swift-storage-0" (UID: "a7f4ca97-1480-4327-8025-9b18a8a7073b") : configmap "swift-ring-files" not found Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.585017 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq5q9\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-kube-api-access-cq5q9\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.591833 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.614473 4678 generic.go:334] "Generic (PLEG): container finished" podID="a755af8b-f01a-4434-9de3-33a059a96ba6" containerID="98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece" exitCode=0 Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.614556 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" event={"ID":"a755af8b-f01a-4434-9de3-33a059a96ba6","Type":"ContainerDied","Data":"98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece"} Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.614586 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" event={"ID":"a755af8b-f01a-4434-9de3-33a059a96ba6","Type":"ContainerDied","Data":"68f57a182a8e6ba95ebdf2404c6dfe213aa4229a454e0d5f3f1840d4c161b21b"} Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.614602 4678 scope.go:117] "RemoveContainer" containerID="98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.614725 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-f947h" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.620176 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" event={"ID":"04529fe4-5999-4649-8dff-674193cd5f1e","Type":"ContainerStarted","Data":"0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a"} Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.620292 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" event={"ID":"04529fe4-5999-4649-8dff-674193cd5f1e","Type":"ContainerStarted","Data":"934b10aa291b7465e3c379fc8fc70d05509fc499db17f30873700c4fda8bbebb"} Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.623608 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"12052ef6-7a13-4ad5-9484-bbad2cabdf5a","Type":"ContainerStarted","Data":"075c9c8c201e793dd10d40ae6ce7ebb4948f57e30d32fb1129ffded6d19d335a"} Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.626445 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"3f6c33f7-7b74-437b-a9fd-3790af864af7","Type":"ContainerStarted","Data":"1a5c197f73f0ac5b8c12884f816ef63c39e0c5287972c4daa3f7dabde1a109c4"} Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.666621 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=21.030175942 podStartE2EDuration="35.666588575s" podCreationTimestamp="2025-12-06 10:53:46 +0000 UTC" firstStartedPulling="2025-12-06 10:54:06.577157659 +0000 UTC m=+1051.420589098" lastFinishedPulling="2025-12-06 10:54:21.213570292 +0000 UTC m=+1066.057001731" observedRunningTime="2025-12-06 10:54:21.661889181 +0000 UTC m=+1066.505320620" watchObservedRunningTime="2025-12-06 10:54:21.666588575 +0000 UTC m=+1066.510020014" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.692913 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.488161721 podStartE2EDuration="35.692891991s" podCreationTimestamp="2025-12-06 10:53:46 +0000 UTC" firstStartedPulling="2025-12-06 10:54:05.94285751 +0000 UTC m=+1050.786288949" lastFinishedPulling="2025-12-06 10:54:21.14758778 +0000 UTC m=+1065.991019219" observedRunningTime="2025-12-06 10:54:21.68731546 +0000 UTC m=+1066.530746899" watchObservedRunningTime="2025-12-06 10:54:21.692891991 +0000 UTC m=+1066.536323430" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.726578 4678 scope.go:117] "RemoveContainer" containerID="0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.735282 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f947h"] Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.742755 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-f947h"] Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.752167 4678 scope.go:117] "RemoveContainer" containerID="98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece" Dec 06 10:54:21 crc kubenswrapper[4678]: E1206 10:54:21.752563 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece\": container with ID starting with 98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece not found: ID does not exist" containerID="98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.752627 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece"} err="failed to get container status \"98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece\": rpc error: code = NotFound desc = could not find container \"98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece\": container with ID starting with 98efeb275a3603e45ebd3ac40742b2cb4972a62b4162c6c4e8aee7f297ea9ece not found: ID does not exist" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.752648 4678 scope.go:117] "RemoveContainer" containerID="0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce" Dec 06 10:54:21 crc kubenswrapper[4678]: E1206 10:54:21.753721 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce\": container with ID starting with 0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce not found: ID does not exist" containerID="0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.753743 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce"} err="failed to get container status \"0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce\": rpc error: code = NotFound desc = could not find container \"0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce\": container with ID starting with 0bea84e3ac45933b1ab122f958a21ca4558fc7e2be50e2864e945d0a906235ce not found: ID does not exist" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.826808 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 06 10:54:21 crc kubenswrapper[4678]: I1206 10:54:21.909770 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 06 10:54:22 crc kubenswrapper[4678]: I1206 10:54:22.071069 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:22 crc kubenswrapper[4678]: E1206 10:54:22.071338 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 10:54:22 crc kubenswrapper[4678]: E1206 10:54:22.071380 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 10:54:22 crc kubenswrapper[4678]: E1206 10:54:22.071464 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift podName:a7f4ca97-1480-4327-8025-9b18a8a7073b nodeName:}" failed. No retries permitted until 2025-12-06 10:54:23.071437353 +0000 UTC m=+1067.914868802 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift") pod "swift-storage-0" (UID: "a7f4ca97-1480-4327-8025-9b18a8a7073b") : configmap "swift-ring-files" not found Dec 06 10:54:22 crc kubenswrapper[4678]: I1206 10:54:22.434646 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 06 10:54:22 crc kubenswrapper[4678]: I1206 10:54:22.625026 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 06 10:54:22 crc kubenswrapper[4678]: I1206 10:54:22.636124 4678 generic.go:334] "Generic (PLEG): container finished" podID="04529fe4-5999-4649-8dff-674193cd5f1e" containerID="0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a" exitCode=0 Dec 06 10:54:22 crc kubenswrapper[4678]: I1206 10:54:22.636204 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" event={"ID":"04529fe4-5999-4649-8dff-674193cd5f1e","Type":"ContainerDied","Data":"0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a"} Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.087444 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:23 crc kubenswrapper[4678]: E1206 10:54:23.087692 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 10:54:23 crc kubenswrapper[4678]: E1206 10:54:23.087717 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 10:54:23 crc kubenswrapper[4678]: E1206 10:54:23.087789 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift podName:a7f4ca97-1480-4327-8025-9b18a8a7073b nodeName:}" failed. No retries permitted until 2025-12-06 10:54:25.087768539 +0000 UTC m=+1069.931199978 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift") pod "swift-storage-0" (UID: "a7f4ca97-1480-4327-8025-9b18a8a7073b") : configmap "swift-ring-files" not found Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.435284 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.452651 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.488057 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a755af8b-f01a-4434-9de3-33a059a96ba6" path="/var/lib/kubelet/pods/a755af8b-f01a-4434-9de3-33a059a96ba6/volumes" Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.489719 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.625480 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.647723 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" event={"ID":"04529fe4-5999-4649-8dff-674193cd5f1e","Type":"ContainerStarted","Data":"a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2"} Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.676601 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" podStartSLOduration=3.676580683 podStartE2EDuration="3.676580683s" podCreationTimestamp="2025-12-06 10:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:23.674932553 +0000 UTC m=+1068.518363992" watchObservedRunningTime="2025-12-06 10:54:23.676580683 +0000 UTC m=+1068.520012122" Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.681334 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 06 10:54:23 crc kubenswrapper[4678]: I1206 10:54:23.727404 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.140144 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-9khdg"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.188676 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-vxq9t"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.193073 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.199921 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.207505 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-vxq9t"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.293602 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-dgln2"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.294896 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.316765 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.317067 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-dgln2"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.318141 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.318179 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg78d\" (UniqueName: \"kubernetes.io/projected/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-kube-api-access-vg78d\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.318277 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-config\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.318308 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-dns-svc\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420004 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-ovn-rundir\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420099 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-config\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420127 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-dns-svc\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420179 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-config\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420205 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420240 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg78d\" (UniqueName: \"kubernetes.io/projected/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-kube-api-access-vg78d\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420269 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-ovs-rundir\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420317 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420337 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs2zz\" (UniqueName: \"kubernetes.io/projected/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-kube-api-access-zs2zz\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.420358 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-combined-ca-bundle\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.421117 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-config\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.421152 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-dns-svc\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.421346 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-ovsdbserver-nb\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.438451 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg78d\" (UniqueName: \"kubernetes.io/projected/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-kube-api-access-vg78d\") pod \"dnsmasq-dns-57d65f699f-vxq9t\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.514384 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.521638 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-config\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.521690 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-ovs-rundir\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.521737 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.521755 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs2zz\" (UniqueName: \"kubernetes.io/projected/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-kube-api-access-zs2zz\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.521771 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-combined-ca-bundle\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.521789 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-ovn-rundir\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.522111 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-ovn-rundir\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.522159 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-ovs-rundir\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.522646 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-config\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.529936 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.540057 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-combined-ca-bundle\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.541826 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs2zz\" (UniqueName: \"kubernetes.io/projected/f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69-kube-api-access-zs2zz\") pod \"ovn-controller-metrics-dgln2\" (UID: \"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69\") " pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.631080 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-dgln2" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.687478 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.803798 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-vxq9t"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.835347 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.838671 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jlb2j"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.840171 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.844820 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.868690 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jlb2j"] Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.937285 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ctvr\" (UniqueName: \"kubernetes.io/projected/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-kube-api-access-8ctvr\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.937392 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.937610 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.937649 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:24 crc kubenswrapper[4678]: I1206 10:54:24.937790 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-config\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.038258 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.038361 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.038379 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.038419 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-config\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.038444 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ctvr\" (UniqueName: \"kubernetes.io/projected/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-kube-api-access-8ctvr\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.039616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.040165 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-config\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.040522 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.040652 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.059531 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ctvr\" (UniqueName: \"kubernetes.io/projected/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-kube-api-access-8ctvr\") pod \"dnsmasq-dns-b8fbc5445-jlb2j\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.079846 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-vxq9t"] Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.139965 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:25 crc kubenswrapper[4678]: E1206 10:54:25.140214 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 10:54:25 crc kubenswrapper[4678]: E1206 10:54:25.140240 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 10:54:25 crc kubenswrapper[4678]: E1206 10:54:25.140294 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift podName:a7f4ca97-1480-4327-8025-9b18a8a7073b nodeName:}" failed. No retries permitted until 2025-12-06 10:54:29.14027589 +0000 UTC m=+1073.983707329 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift") pod "swift-storage-0" (UID: "a7f4ca97-1480-4327-8025-9b18a8a7073b") : configmap "swift-ring-files" not found Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.197400 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.198880 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.203823 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-z788f" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.203985 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.204158 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.205644 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.211356 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.226220 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.288622 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-dgln2"] Dec 06 10:54:25 crc kubenswrapper[4678]: W1206 10:54:25.306976 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7dd8a89_6c2f_4dcf_96ed_1e9a04bdec69.slice/crio-0135a083719e59e0e67fc5bd5d25ad7ea64404e4f63e516f8f856f2c9f8cf7c3 WatchSource:0}: Error finding container 0135a083719e59e0e67fc5bd5d25ad7ea64404e4f63e516f8f856f2c9f8cf7c3: Status 404 returned error can't find the container with id 0135a083719e59e0e67fc5bd5d25ad7ea64404e4f63e516f8f856f2c9f8cf7c3 Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.309506 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-tkg52"] Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.310899 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.315652 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.315986 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.316193 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.334470 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tkg52"] Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.345587 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.345636 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.345667 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5a586aa5-fe85-49da-a10e-3e29de7dba4d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.345693 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a586aa5-fe85-49da-a10e-3e29de7dba4d-config\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.346023 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a586aa5-fe85-49da-a10e-3e29de7dba4d-scripts\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.346060 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.346094 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfmj8\" (UniqueName: \"kubernetes.io/projected/5a586aa5-fe85-49da-a10e-3e29de7dba4d-kube-api-access-mfmj8\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448022 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448085 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448130 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5a586aa5-fe85-49da-a10e-3e29de7dba4d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448165 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a586aa5-fe85-49da-a10e-3e29de7dba4d-config\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448192 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-dispersionconf\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448218 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-swiftconf\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448274 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-scripts\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448319 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a586aa5-fe85-49da-a10e-3e29de7dba4d-scripts\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448348 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448371 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkvq5\" (UniqueName: \"kubernetes.io/projected/db2561ff-632a-463d-a7af-79d83536812b-kube-api-access-bkvq5\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448403 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-ring-data-devices\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.448438 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfmj8\" (UniqueName: \"kubernetes.io/projected/5a586aa5-fe85-49da-a10e-3e29de7dba4d-kube-api-access-mfmj8\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.449065 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-combined-ca-bundle\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.449105 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2561ff-632a-463d-a7af-79d83536812b-etc-swift\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.449374 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5a586aa5-fe85-49da-a10e-3e29de7dba4d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.450371 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a586aa5-fe85-49da-a10e-3e29de7dba4d-scripts\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.450711 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5a586aa5-fe85-49da-a10e-3e29de7dba4d-config\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.454572 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.465515 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.465822 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a586aa5-fe85-49da-a10e-3e29de7dba4d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.469234 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfmj8\" (UniqueName: \"kubernetes.io/projected/5a586aa5-fe85-49da-a10e-3e29de7dba4d-kube-api-access-mfmj8\") pod \"ovn-northd-0\" (UID: \"5a586aa5-fe85-49da-a10e-3e29de7dba4d\") " pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.549350 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.577320 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-combined-ca-bundle\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.577724 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2561ff-632a-463d-a7af-79d83536812b-etc-swift\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.577880 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-dispersionconf\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.579187 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-swiftconf\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.579295 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-scripts\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.579382 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkvq5\" (UniqueName: \"kubernetes.io/projected/db2561ff-632a-463d-a7af-79d83536812b-kube-api-access-bkvq5\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.579417 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-ring-data-devices\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.596992 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2561ff-632a-463d-a7af-79d83536812b-etc-swift\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.599191 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-scripts\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.601535 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-ring-data-devices\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.612005 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-swiftconf\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.618240 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-dispersionconf\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.649503 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-combined-ca-bundle\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.653300 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkvq5\" (UniqueName: \"kubernetes.io/projected/db2561ff-632a-463d-a7af-79d83536812b-kube-api-access-bkvq5\") pod \"swift-ring-rebalance-tkg52\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.677858 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jlb2j"] Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.695707 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.737797 4678 generic.go:334] "Generic (PLEG): container finished" podID="fc1af712-5342-4157-b7b6-eb9d5aa69bd7" containerID="781bd5e7b3492c1162770e9cdb38bf0d5dc1f9971c0585a7b47ce26b86200c31" exitCode=0 Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.737836 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" event={"ID":"fc1af712-5342-4157-b7b6-eb9d5aa69bd7","Type":"ContainerDied","Data":"781bd5e7b3492c1162770e9cdb38bf0d5dc1f9971c0585a7b47ce26b86200c31"} Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.737870 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" event={"ID":"fc1af712-5342-4157-b7b6-eb9d5aa69bd7","Type":"ContainerStarted","Data":"006daee69581ff674dafa3cfee2222eb914037ba924060734a80d590bbf0d0e6"} Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.740921 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-dgln2" event={"ID":"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69","Type":"ContainerStarted","Data":"0135a083719e59e0e67fc5bd5d25ad7ea64404e4f63e516f8f856f2c9f8cf7c3"} Dec 06 10:54:25 crc kubenswrapper[4678]: I1206 10:54:25.741636 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" podUID="04529fe4-5999-4649-8dff-674193cd5f1e" containerName="dnsmasq-dns" containerID="cri-o://a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2" gracePeriod=10 Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.194994 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.199137 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 06 10:54:26 crc kubenswrapper[4678]: W1206 10:54:26.201797 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a586aa5_fe85_49da_a10e_3e29de7dba4d.slice/crio-f7c1b482da62e30eb632a6c47e2974dce96c62f3735728f178f92f36013ba475 WatchSource:0}: Error finding container f7c1b482da62e30eb632a6c47e2974dce96c62f3735728f178f92f36013ba475: Status 404 returned error can't find the container with id f7c1b482da62e30eb632a6c47e2974dce96c62f3735728f178f92f36013ba475 Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.316704 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vg78d\" (UniqueName: \"kubernetes.io/projected/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-kube-api-access-vg78d\") pod \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.316815 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-ovsdbserver-nb\") pod \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.316954 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-dns-svc\") pod \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.317004 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-config\") pod \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\" (UID: \"fc1af712-5342-4157-b7b6-eb9d5aa69bd7\") " Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.323515 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-kube-api-access-vg78d" (OuterVolumeSpecName: "kube-api-access-vg78d") pod "fc1af712-5342-4157-b7b6-eb9d5aa69bd7" (UID: "fc1af712-5342-4157-b7b6-eb9d5aa69bd7"). InnerVolumeSpecName "kube-api-access-vg78d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.366316 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fc1af712-5342-4157-b7b6-eb9d5aa69bd7" (UID: "fc1af712-5342-4157-b7b6-eb9d5aa69bd7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.366831 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-config" (OuterVolumeSpecName: "config") pod "fc1af712-5342-4157-b7b6-eb9d5aa69bd7" (UID: "fc1af712-5342-4157-b7b6-eb9d5aa69bd7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.375025 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc1af712-5342-4157-b7b6-eb9d5aa69bd7" (UID: "fc1af712-5342-4157-b7b6-eb9d5aa69bd7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.401540 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.419448 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.419496 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.419509 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vg78d\" (UniqueName: \"kubernetes.io/projected/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-kube-api-access-vg78d\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.419521 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc1af712-5342-4157-b7b6-eb9d5aa69bd7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.503311 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-tkg52"] Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.520185 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-dns-svc\") pod \"04529fe4-5999-4649-8dff-674193cd5f1e\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.520349 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9b7zw\" (UniqueName: \"kubernetes.io/projected/04529fe4-5999-4649-8dff-674193cd5f1e-kube-api-access-9b7zw\") pod \"04529fe4-5999-4649-8dff-674193cd5f1e\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.520404 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-config\") pod \"04529fe4-5999-4649-8dff-674193cd5f1e\" (UID: \"04529fe4-5999-4649-8dff-674193cd5f1e\") " Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.534958 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04529fe4-5999-4649-8dff-674193cd5f1e-kube-api-access-9b7zw" (OuterVolumeSpecName: "kube-api-access-9b7zw") pod "04529fe4-5999-4649-8dff-674193cd5f1e" (UID: "04529fe4-5999-4649-8dff-674193cd5f1e"). InnerVolumeSpecName "kube-api-access-9b7zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.566637 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-config" (OuterVolumeSpecName: "config") pod "04529fe4-5999-4649-8dff-674193cd5f1e" (UID: "04529fe4-5999-4649-8dff-674193cd5f1e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.611228 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "04529fe4-5999-4649-8dff-674193cd5f1e" (UID: "04529fe4-5999-4649-8dff-674193cd5f1e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.622792 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.622822 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9b7zw\" (UniqueName: \"kubernetes.io/projected/04529fe4-5999-4649-8dff-674193cd5f1e-kube-api-access-9b7zw\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.622831 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04529fe4-5999-4649-8dff-674193cd5f1e-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.664509 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.665083 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.741390 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.750385 4678 generic.go:334] "Generic (PLEG): container finished" podID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerID="e8aa9719bf55fc0c42d1c39280635cfceb8166de99c0a5f8a8c0ee0ae11053f7" exitCode=0 Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.750451 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" event={"ID":"4c396d76-0ede-4141-b9be-ffba8a8f8f2d","Type":"ContainerDied","Data":"e8aa9719bf55fc0c42d1c39280635cfceb8166de99c0a5f8a8c0ee0ae11053f7"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.750476 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" event={"ID":"4c396d76-0ede-4141-b9be-ffba8a8f8f2d","Type":"ContainerStarted","Data":"3de61fa0da586a30c3a305cb0d5c57592fdd92000e46c664de96d777d112ce4f"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.753745 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tkg52" event={"ID":"db2561ff-632a-463d-a7af-79d83536812b","Type":"ContainerStarted","Data":"220d7d9e72986693b1b6bd63085e832e17654e1c1f86b1f22b155321d1997150"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.755944 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-dgln2" event={"ID":"f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69","Type":"ContainerStarted","Data":"7a6240b98a1e6c482957ee9c3fa1244c3760a2dc97f1496bc525bcf753eb8294"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.757225 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5a586aa5-fe85-49da-a10e-3e29de7dba4d","Type":"ContainerStarted","Data":"f7c1b482da62e30eb632a6c47e2974dce96c62f3735728f178f92f36013ba475"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.759165 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.759160 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d65f699f-vxq9t" event={"ID":"fc1af712-5342-4157-b7b6-eb9d5aa69bd7","Type":"ContainerDied","Data":"006daee69581ff674dafa3cfee2222eb914037ba924060734a80d590bbf0d0e6"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.759275 4678 scope.go:117] "RemoveContainer" containerID="781bd5e7b3492c1162770e9cdb38bf0d5dc1f9971c0585a7b47ce26b86200c31" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.772616 4678 generic.go:334] "Generic (PLEG): container finished" podID="04529fe4-5999-4649-8dff-674193cd5f1e" containerID="a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2" exitCode=0 Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.773375 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.774677 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" event={"ID":"04529fe4-5999-4649-8dff-674193cd5f1e","Type":"ContainerDied","Data":"a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.774730 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-9khdg" event={"ID":"04529fe4-5999-4649-8dff-674193cd5f1e","Type":"ContainerDied","Data":"934b10aa291b7465e3c379fc8fc70d05509fc499db17f30873700c4fda8bbebb"} Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.784325 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-dgln2" podStartSLOduration=2.784302313 podStartE2EDuration="2.784302313s" podCreationTimestamp="2025-12-06 10:54:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:26.782413106 +0000 UTC m=+1071.625844545" watchObservedRunningTime="2025-12-06 10:54:26.784302313 +0000 UTC m=+1071.627733752" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.816517 4678 scope.go:117] "RemoveContainer" containerID="a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.858652 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-vxq9t"] Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.880135 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d65f699f-vxq9t"] Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.893615 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-9khdg"] Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.898417 4678 scope.go:117] "RemoveContainer" containerID="0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.899689 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-9khdg"] Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.967204 4678 scope.go:117] "RemoveContainer" containerID="a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2" Dec 06 10:54:26 crc kubenswrapper[4678]: E1206 10:54:26.967887 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2\": container with ID starting with a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2 not found: ID does not exist" containerID="a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.967948 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2"} err="failed to get container status \"a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2\": rpc error: code = NotFound desc = could not find container \"a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2\": container with ID starting with a193859da69a9af905cbbfadbab4562e8d398a6e2eeb7578ffd0ad2347f9d3c2 not found: ID does not exist" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.967976 4678 scope.go:117] "RemoveContainer" containerID="0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a" Dec 06 10:54:26 crc kubenswrapper[4678]: E1206 10:54:26.968305 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a\": container with ID starting with 0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a not found: ID does not exist" containerID="0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a" Dec 06 10:54:26 crc kubenswrapper[4678]: I1206 10:54:26.968352 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a"} err="failed to get container status \"0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a\": rpc error: code = NotFound desc = could not find container \"0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a\": container with ID starting with 0c792142621b1ced07ac3acde67295bedd472f59f8c8a5b26385643c02205e2a not found: ID does not exist" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.012503 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.504605 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04529fe4-5999-4649-8dff-674193cd5f1e" path="/var/lib/kubelet/pods/04529fe4-5999-4649-8dff-674193cd5f1e/volumes" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.505589 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc1af712-5342-4157-b7b6-eb9d5aa69bd7" path="/var/lib/kubelet/pods/fc1af712-5342-4157-b7b6-eb9d5aa69bd7/volumes" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.725650 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-mwspd"] Dec 06 10:54:27 crc kubenswrapper[4678]: E1206 10:54:27.726026 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04529fe4-5999-4649-8dff-674193cd5f1e" containerName="init" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.726044 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="04529fe4-5999-4649-8dff-674193cd5f1e" containerName="init" Dec 06 10:54:27 crc kubenswrapper[4678]: E1206 10:54:27.726068 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1af712-5342-4157-b7b6-eb9d5aa69bd7" containerName="init" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.726074 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1af712-5342-4157-b7b6-eb9d5aa69bd7" containerName="init" Dec 06 10:54:27 crc kubenswrapper[4678]: E1206 10:54:27.726092 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04529fe4-5999-4649-8dff-674193cd5f1e" containerName="dnsmasq-dns" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.726108 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="04529fe4-5999-4649-8dff-674193cd5f1e" containerName="dnsmasq-dns" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.726258 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="04529fe4-5999-4649-8dff-674193cd5f1e" containerName="dnsmasq-dns" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.726273 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1af712-5342-4157-b7b6-eb9d5aa69bd7" containerName="init" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.726849 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.734134 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mwspd"] Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.789939 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" event={"ID":"4c396d76-0ede-4141-b9be-ffba8a8f8f2d","Type":"ContainerStarted","Data":"ae64fa013429922acaccb6ecd8a497f70287113248adc2fc9df366cb096103c2"} Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.790004 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.828999 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-e1f4-account-create-update-pnscz"] Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.830438 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.832126 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" podStartSLOduration=3.832104565 podStartE2EDuration="3.832104565s" podCreationTimestamp="2025-12-06 10:54:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:27.825594235 +0000 UTC m=+1072.669025664" watchObservedRunningTime="2025-12-06 10:54:27.832104565 +0000 UTC m=+1072.675536024" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.835612 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.844709 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e1f4-account-create-update-pnscz"] Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.860391 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-operator-scripts\") pod \"keystone-db-create-mwspd\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.860576 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4m94s\" (UniqueName: \"kubernetes.io/projected/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-kube-api-access-4m94s\") pod \"keystone-db-create-mwspd\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.962415 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtk9b\" (UniqueName: \"kubernetes.io/projected/f09ba32d-29da-4907-80b0-4df3433704f8-kube-api-access-vtk9b\") pod \"keystone-e1f4-account-create-update-pnscz\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.962468 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4m94s\" (UniqueName: \"kubernetes.io/projected/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-kube-api-access-4m94s\") pod \"keystone-db-create-mwspd\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.962619 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-operator-scripts\") pod \"keystone-db-create-mwspd\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.962655 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f09ba32d-29da-4907-80b0-4df3433704f8-operator-scripts\") pod \"keystone-e1f4-account-create-update-pnscz\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.966924 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-operator-scripts\") pod \"keystone-db-create-mwspd\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:27 crc kubenswrapper[4678]: I1206 10:54:27.981308 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4m94s\" (UniqueName: \"kubernetes.io/projected/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-kube-api-access-4m94s\") pod \"keystone-db-create-mwspd\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.041297 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.064470 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f09ba32d-29da-4907-80b0-4df3433704f8-operator-scripts\") pod \"keystone-e1f4-account-create-update-pnscz\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.064890 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtk9b\" (UniqueName: \"kubernetes.io/projected/f09ba32d-29da-4907-80b0-4df3433704f8-kube-api-access-vtk9b\") pod \"keystone-e1f4-account-create-update-pnscz\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.066115 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f09ba32d-29da-4907-80b0-4df3433704f8-operator-scripts\") pod \"keystone-e1f4-account-create-update-pnscz\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.090168 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-5mckr"] Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.091241 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.093405 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtk9b\" (UniqueName: \"kubernetes.io/projected/f09ba32d-29da-4907-80b0-4df3433704f8-kube-api-access-vtk9b\") pod \"keystone-e1f4-account-create-update-pnscz\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.116055 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5mckr"] Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.147855 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.166149 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-17f0-account-create-update-8xfqp"] Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.166164 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddf2w\" (UniqueName: \"kubernetes.io/projected/2f077be8-3497-464c-9bfd-4752c6bc08c3-kube-api-access-ddf2w\") pod \"placement-db-create-5mckr\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.166406 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f077be8-3497-464c-9bfd-4752c6bc08c3-operator-scripts\") pod \"placement-db-create-5mckr\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.167266 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.171535 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.175156 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-17f0-account-create-update-8xfqp"] Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.268605 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f077be8-3497-464c-9bfd-4752c6bc08c3-operator-scripts\") pod \"placement-db-create-5mckr\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.268875 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snfkf\" (UniqueName: \"kubernetes.io/projected/b226f5f5-03ea-4614-bd24-5499b6fe7457-kube-api-access-snfkf\") pod \"placement-17f0-account-create-update-8xfqp\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.269005 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b226f5f5-03ea-4614-bd24-5499b6fe7457-operator-scripts\") pod \"placement-17f0-account-create-update-8xfqp\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.269039 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddf2w\" (UniqueName: \"kubernetes.io/projected/2f077be8-3497-464c-9bfd-4752c6bc08c3-kube-api-access-ddf2w\") pod \"placement-db-create-5mckr\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.269737 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f077be8-3497-464c-9bfd-4752c6bc08c3-operator-scripts\") pod \"placement-db-create-5mckr\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.291710 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddf2w\" (UniqueName: \"kubernetes.io/projected/2f077be8-3497-464c-9bfd-4752c6bc08c3-kube-api-access-ddf2w\") pod \"placement-db-create-5mckr\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.370980 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b226f5f5-03ea-4614-bd24-5499b6fe7457-operator-scripts\") pod \"placement-17f0-account-create-update-8xfqp\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.371108 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snfkf\" (UniqueName: \"kubernetes.io/projected/b226f5f5-03ea-4614-bd24-5499b6fe7457-kube-api-access-snfkf\") pod \"placement-17f0-account-create-update-8xfqp\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.372377 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b226f5f5-03ea-4614-bd24-5499b6fe7457-operator-scripts\") pod \"placement-17f0-account-create-update-8xfqp\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.401163 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snfkf\" (UniqueName: \"kubernetes.io/projected/b226f5f5-03ea-4614-bd24-5499b6fe7457-kube-api-access-snfkf\") pod \"placement-17f0-account-create-update-8xfqp\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.453010 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mckr" Dec 06 10:54:28 crc kubenswrapper[4678]: I1206 10:54:28.490305 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:29 crc kubenswrapper[4678]: I1206 10:54:29.185381 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:29 crc kubenswrapper[4678]: E1206 10:54:29.185798 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 10:54:29 crc kubenswrapper[4678]: E1206 10:54:29.185830 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 10:54:29 crc kubenswrapper[4678]: E1206 10:54:29.185949 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift podName:a7f4ca97-1480-4327-8025-9b18a8a7073b nodeName:}" failed. No retries permitted until 2025-12-06 10:54:37.185883163 +0000 UTC m=+1082.029314602 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift") pod "swift-storage-0" (UID: "a7f4ca97-1480-4327-8025-9b18a8a7073b") : configmap "swift-ring-files" not found Dec 06 10:54:30 crc kubenswrapper[4678]: I1206 10:54:30.771843 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.306540 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-xvjr7"] Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.307977 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.321234 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xvjr7"] Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.352186 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f8e528-9dfd-4572-9ad6-07d8393be1bb-operator-scripts\") pod \"glance-db-create-xvjr7\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.352281 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s2j6\" (UniqueName: \"kubernetes.io/projected/51f8e528-9dfd-4572-9ad6-07d8393be1bb-kube-api-access-5s2j6\") pod \"glance-db-create-xvjr7\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.440627 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-98b3-account-create-update-4rg7b"] Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.442053 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.445351 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.453462 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5efdf59-1cdd-4aef-8105-f44519426d2c-operator-scripts\") pod \"glance-98b3-account-create-update-4rg7b\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.453553 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k5rl\" (UniqueName: \"kubernetes.io/projected/c5efdf59-1cdd-4aef-8105-f44519426d2c-kube-api-access-2k5rl\") pod \"glance-98b3-account-create-update-4rg7b\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.453588 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f8e528-9dfd-4572-9ad6-07d8393be1bb-operator-scripts\") pod \"glance-db-create-xvjr7\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.453747 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s2j6\" (UniqueName: \"kubernetes.io/projected/51f8e528-9dfd-4572-9ad6-07d8393be1bb-kube-api-access-5s2j6\") pod \"glance-db-create-xvjr7\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.454390 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f8e528-9dfd-4572-9ad6-07d8393be1bb-operator-scripts\") pod \"glance-db-create-xvjr7\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.455770 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-98b3-account-create-update-4rg7b"] Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.483931 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s2j6\" (UniqueName: \"kubernetes.io/projected/51f8e528-9dfd-4572-9ad6-07d8393be1bb-kube-api-access-5s2j6\") pod \"glance-db-create-xvjr7\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.554738 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5efdf59-1cdd-4aef-8105-f44519426d2c-operator-scripts\") pod \"glance-98b3-account-create-update-4rg7b\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.555716 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k5rl\" (UniqueName: \"kubernetes.io/projected/c5efdf59-1cdd-4aef-8105-f44519426d2c-kube-api-access-2k5rl\") pod \"glance-98b3-account-create-update-4rg7b\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.558197 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5efdf59-1cdd-4aef-8105-f44519426d2c-operator-scripts\") pod \"glance-98b3-account-create-update-4rg7b\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.572694 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k5rl\" (UniqueName: \"kubernetes.io/projected/c5efdf59-1cdd-4aef-8105-f44519426d2c-kube-api-access-2k5rl\") pod \"glance-98b3-account-create-update-4rg7b\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.630153 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:33 crc kubenswrapper[4678]: I1206 10:54:33.756838 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.407680 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-xvjr7"] Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.443941 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-17f0-account-create-update-8xfqp"] Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.450505 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mwspd"] Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.575479 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-e1f4-account-create-update-pnscz"] Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.590961 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-5mckr"] Dec 06 10:54:34 crc kubenswrapper[4678]: W1206 10:54:34.601463 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5efdf59_1cdd_4aef_8105_f44519426d2c.slice/crio-9c38336d62298d2291ced67070680b1c5951e3a76e9ae53238c5eefa730ed01c WatchSource:0}: Error finding container 9c38336d62298d2291ced67070680b1c5951e3a76e9ae53238c5eefa730ed01c: Status 404 returned error can't find the container with id 9c38336d62298d2291ced67070680b1c5951e3a76e9ae53238c5eefa730ed01c Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.608622 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-98b3-account-create-update-4rg7b"] Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.908509 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tkg52" event={"ID":"db2561ff-632a-463d-a7af-79d83536812b","Type":"ContainerStarted","Data":"365e9887f82f39fbb4167c9c235cf43f46b2de75ac60f2bbd15d0bfa722bece5"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.909828 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-17f0-account-create-update-8xfqp" event={"ID":"b226f5f5-03ea-4614-bd24-5499b6fe7457","Type":"ContainerStarted","Data":"2d4260b04317a574b7d3e38bee42021cf319114d7ce64e95ac732240971fe18c"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.909857 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-17f0-account-create-update-8xfqp" event={"ID":"b226f5f5-03ea-4614-bd24-5499b6fe7457","Type":"ContainerStarted","Data":"471d2e62258aea5c12e18a7f7b2b058fe0714c0f16d2f240ee3d7f2e4574256d"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.912946 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-98b3-account-create-update-4rg7b" event={"ID":"c5efdf59-1cdd-4aef-8105-f44519426d2c","Type":"ContainerStarted","Data":"05d55e96ccf43e190ec08bef298678e3ac341df89d7b5503b74c1dd34bf249be"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.912988 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-98b3-account-create-update-4rg7b" event={"ID":"c5efdf59-1cdd-4aef-8105-f44519426d2c","Type":"ContainerStarted","Data":"9c38336d62298d2291ced67070680b1c5951e3a76e9ae53238c5eefa730ed01c"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.915796 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mckr" event={"ID":"2f077be8-3497-464c-9bfd-4752c6bc08c3","Type":"ContainerStarted","Data":"e5984f9e8537e9c8f3f4248f254aed3cad0f0398bce3fb3abfea5471cda418e5"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.915828 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mckr" event={"ID":"2f077be8-3497-464c-9bfd-4752c6bc08c3","Type":"ContainerStarted","Data":"7e2ea59d24fe0530cc81357c2d22c19c29d21e1d08148a171e0a99d4d768f541"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.918455 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5a586aa5-fe85-49da-a10e-3e29de7dba4d","Type":"ContainerStarted","Data":"8406991b86320b9299c4864420c5f23d7439bfe27afbe9d638f3ef104da88e98"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.918508 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"5a586aa5-fe85-49da-a10e-3e29de7dba4d","Type":"ContainerStarted","Data":"04f69fbd8641e9937707f2aa8cc230a24c68c1693d184882f9ae98badf80a9e3"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.919139 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.920642 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mwspd" event={"ID":"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b","Type":"ContainerStarted","Data":"b0a5cbee3899208a14d1322cbf9a30eb3f16965598f6873e3e7e9e9195d57916"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.920668 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mwspd" event={"ID":"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b","Type":"ContainerStarted","Data":"cdda316e3e748e03aa9ff169e021006f1c02d4d05306e381c9a8286814025ea9"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.923292 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xvjr7" event={"ID":"51f8e528-9dfd-4572-9ad6-07d8393be1bb","Type":"ContainerStarted","Data":"5c1b064acf9cadbd8425c6c46461651efbbe9097287298ae7d003d9ce0411c33"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.923340 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xvjr7" event={"ID":"51f8e528-9dfd-4572-9ad6-07d8393be1bb","Type":"ContainerStarted","Data":"43b7d2d156f6b9165d38639ffa0cbaa5a08a1b1c75de6490393670316fdb9958"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.928421 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e1f4-account-create-update-pnscz" event={"ID":"f09ba32d-29da-4907-80b0-4df3433704f8","Type":"ContainerStarted","Data":"60408b295cd3659ac8b9add1dcd546a11edddc50c9450b186c1a2d352c6434e3"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.928476 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e1f4-account-create-update-pnscz" event={"ID":"f09ba32d-29da-4907-80b0-4df3433704f8","Type":"ContainerStarted","Data":"7070518f1bdbbcdae7c2f5693177bf2b3119eb8775e3a9a07de4db877b303ada"} Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.930645 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-tkg52" podStartSLOduration=2.6672572409999997 podStartE2EDuration="9.930625107s" podCreationTimestamp="2025-12-06 10:54:25 +0000 UTC" firstStartedPulling="2025-12-06 10:54:26.520942742 +0000 UTC m=+1071.364374181" lastFinishedPulling="2025-12-06 10:54:33.784310598 +0000 UTC m=+1078.627742047" observedRunningTime="2025-12-06 10:54:34.925726568 +0000 UTC m=+1079.769158007" watchObservedRunningTime="2025-12-06 10:54:34.930625107 +0000 UTC m=+1079.774056546" Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.944343 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-98b3-account-create-update-4rg7b" podStartSLOduration=1.944322498 podStartE2EDuration="1.944322498s" podCreationTimestamp="2025-12-06 10:54:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:34.938887871 +0000 UTC m=+1079.782319310" watchObservedRunningTime="2025-12-06 10:54:34.944322498 +0000 UTC m=+1079.787753937" Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.957093 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-xvjr7" podStartSLOduration=1.957069408 podStartE2EDuration="1.957069408s" podCreationTimestamp="2025-12-06 10:54:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:34.953572051 +0000 UTC m=+1079.797003490" watchObservedRunningTime="2025-12-06 10:54:34.957069408 +0000 UTC m=+1079.800500837" Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.970337 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-mwspd" podStartSLOduration=7.96986738 podStartE2EDuration="7.96986738s" podCreationTimestamp="2025-12-06 10:54:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:34.968438517 +0000 UTC m=+1079.811869956" watchObservedRunningTime="2025-12-06 10:54:34.96986738 +0000 UTC m=+1079.813298819" Dec 06 10:54:34 crc kubenswrapper[4678]: I1206 10:54:34.996520 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.538825447 podStartE2EDuration="9.996482236s" podCreationTimestamp="2025-12-06 10:54:25 +0000 UTC" firstStartedPulling="2025-12-06 10:54:26.205427333 +0000 UTC m=+1071.048858772" lastFinishedPulling="2025-12-06 10:54:33.663084122 +0000 UTC m=+1078.506515561" observedRunningTime="2025-12-06 10:54:34.989957506 +0000 UTC m=+1079.833388945" watchObservedRunningTime="2025-12-06 10:54:34.996482236 +0000 UTC m=+1079.839913675" Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.008431 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-5mckr" podStartSLOduration=7.008408262 podStartE2EDuration="7.008408262s" podCreationTimestamp="2025-12-06 10:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:35.002926834 +0000 UTC m=+1079.846358273" watchObservedRunningTime="2025-12-06 10:54:35.008408262 +0000 UTC m=+1079.851839701" Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.021879 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-17f0-account-create-update-8xfqp" podStartSLOduration=7.021858724 podStartE2EDuration="7.021858724s" podCreationTimestamp="2025-12-06 10:54:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:35.018257603 +0000 UTC m=+1079.861689042" watchObservedRunningTime="2025-12-06 10:54:35.021858724 +0000 UTC m=+1079.865290163" Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.036940 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-e1f4-account-create-update-pnscz" podStartSLOduration=8.036916425 podStartE2EDuration="8.036916425s" podCreationTimestamp="2025-12-06 10:54:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:35.034532722 +0000 UTC m=+1079.877964161" watchObservedRunningTime="2025-12-06 10:54:35.036916425 +0000 UTC m=+1079.880347864" Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.212729 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.287551 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fdnth"] Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.287777 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerName="dnsmasq-dns" containerID="cri-o://2c94b4c0a6c2dbabaf6944ff038082789e635ac8c153e525d547a63b452a54bd" gracePeriod=10 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.940054 4678 generic.go:334] "Generic (PLEG): container finished" podID="e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b" containerID="b0a5cbee3899208a14d1322cbf9a30eb3f16965598f6873e3e7e9e9195d57916" exitCode=0 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.940383 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mwspd" event={"ID":"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b","Type":"ContainerDied","Data":"b0a5cbee3899208a14d1322cbf9a30eb3f16965598f6873e3e7e9e9195d57916"} Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.946568 4678 generic.go:334] "Generic (PLEG): container finished" podID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerID="2c94b4c0a6c2dbabaf6944ff038082789e635ac8c153e525d547a63b452a54bd" exitCode=0 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.946633 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" event={"ID":"9ac16ad4-74b1-4968-bd23-10c814ed087d","Type":"ContainerDied","Data":"2c94b4c0a6c2dbabaf6944ff038082789e635ac8c153e525d547a63b452a54bd"} Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.946660 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" event={"ID":"9ac16ad4-74b1-4968-bd23-10c814ed087d","Type":"ContainerDied","Data":"adda73ac535542f65b1d15ddafe2f9fde1f05c91196cdaaf5079600f85f678a2"} Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.946671 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adda73ac535542f65b1d15ddafe2f9fde1f05c91196cdaaf5079600f85f678a2" Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.948939 4678 generic.go:334] "Generic (PLEG): container finished" podID="51f8e528-9dfd-4572-9ad6-07d8393be1bb" containerID="5c1b064acf9cadbd8425c6c46461651efbbe9097287298ae7d003d9ce0411c33" exitCode=0 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.948998 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xvjr7" event={"ID":"51f8e528-9dfd-4572-9ad6-07d8393be1bb","Type":"ContainerDied","Data":"5c1b064acf9cadbd8425c6c46461651efbbe9097287298ae7d003d9ce0411c33"} Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.951860 4678 generic.go:334] "Generic (PLEG): container finished" podID="f09ba32d-29da-4907-80b0-4df3433704f8" containerID="60408b295cd3659ac8b9add1dcd546a11edddc50c9450b186c1a2d352c6434e3" exitCode=0 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.951999 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e1f4-account-create-update-pnscz" event={"ID":"f09ba32d-29da-4907-80b0-4df3433704f8","Type":"ContainerDied","Data":"60408b295cd3659ac8b9add1dcd546a11edddc50c9450b186c1a2d352c6434e3"} Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.964851 4678 generic.go:334] "Generic (PLEG): container finished" podID="b226f5f5-03ea-4614-bd24-5499b6fe7457" containerID="2d4260b04317a574b7d3e38bee42021cf319114d7ce64e95ac732240971fe18c" exitCode=0 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.964919 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-17f0-account-create-update-8xfqp" event={"ID":"b226f5f5-03ea-4614-bd24-5499b6fe7457","Type":"ContainerDied","Data":"2d4260b04317a574b7d3e38bee42021cf319114d7ce64e95ac732240971fe18c"} Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.969094 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.969894 4678 generic.go:334] "Generic (PLEG): container finished" podID="c5efdf59-1cdd-4aef-8105-f44519426d2c" containerID="05d55e96ccf43e190ec08bef298678e3ac341df89d7b5503b74c1dd34bf249be" exitCode=0 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.969978 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-98b3-account-create-update-4rg7b" event={"ID":"c5efdf59-1cdd-4aef-8105-f44519426d2c","Type":"ContainerDied","Data":"05d55e96ccf43e190ec08bef298678e3ac341df89d7b5503b74c1dd34bf249be"} Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.971256 4678 generic.go:334] "Generic (PLEG): container finished" podID="2f077be8-3497-464c-9bfd-4752c6bc08c3" containerID="e5984f9e8537e9c8f3f4248f254aed3cad0f0398bce3fb3abfea5471cda418e5" exitCode=0 Dec 06 10:54:35 crc kubenswrapper[4678]: I1206 10:54:35.972152 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mckr" event={"ID":"2f077be8-3497-464c-9bfd-4752c6bc08c3","Type":"ContainerDied","Data":"e5984f9e8537e9c8f3f4248f254aed3cad0f0398bce3fb3abfea5471cda418e5"} Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.041814 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-config\") pod \"9ac16ad4-74b1-4968-bd23-10c814ed087d\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.041947 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf8rq\" (UniqueName: \"kubernetes.io/projected/9ac16ad4-74b1-4968-bd23-10c814ed087d-kube-api-access-rf8rq\") pod \"9ac16ad4-74b1-4968-bd23-10c814ed087d\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.042056 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-dns-svc\") pod \"9ac16ad4-74b1-4968-bd23-10c814ed087d\" (UID: \"9ac16ad4-74b1-4968-bd23-10c814ed087d\") " Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.055776 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac16ad4-74b1-4968-bd23-10c814ed087d-kube-api-access-rf8rq" (OuterVolumeSpecName: "kube-api-access-rf8rq") pod "9ac16ad4-74b1-4968-bd23-10c814ed087d" (UID: "9ac16ad4-74b1-4968-bd23-10c814ed087d"). InnerVolumeSpecName "kube-api-access-rf8rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.146183 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf8rq\" (UniqueName: \"kubernetes.io/projected/9ac16ad4-74b1-4968-bd23-10c814ed087d-kube-api-access-rf8rq\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.151909 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ac16ad4-74b1-4968-bd23-10c814ed087d" (UID: "9ac16ad4-74b1-4968-bd23-10c814ed087d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.160816 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-config" (OuterVolumeSpecName: "config") pod "9ac16ad4-74b1-4968-bd23-10c814ed087d" (UID: "9ac16ad4-74b1-4968-bd23-10c814ed087d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.248066 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.248103 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac16ad4-74b1-4968-bd23-10c814ed087d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:36 crc kubenswrapper[4678]: I1206 10:54:36.977940 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-fdnth" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.022099 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fdnth"] Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.027670 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-fdnth"] Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.263455 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:37 crc kubenswrapper[4678]: E1206 10:54:37.263806 4678 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 06 10:54:37 crc kubenswrapper[4678]: E1206 10:54:37.263830 4678 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 06 10:54:37 crc kubenswrapper[4678]: E1206 10:54:37.263887 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift podName:a7f4ca97-1480-4327-8025-9b18a8a7073b nodeName:}" failed. No retries permitted until 2025-12-06 10:54:53.263869653 +0000 UTC m=+1098.107301092 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift") pod "swift-storage-0" (UID: "a7f4ca97-1480-4327-8025-9b18a8a7073b") : configmap "swift-ring-files" not found Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.404514 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.492751 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" path="/var/lib/kubelet/pods/9ac16ad4-74b1-4968-bd23-10c814ed087d/volumes" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.573203 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4m94s\" (UniqueName: \"kubernetes.io/projected/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-kube-api-access-4m94s\") pod \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.573361 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-operator-scripts\") pod \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\" (UID: \"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.576451 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b" (UID: "e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.582527 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-kube-api-access-4m94s" (OuterVolumeSpecName: "kube-api-access-4m94s") pod "e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b" (UID: "e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b"). InnerVolumeSpecName "kube-api-access-4m94s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.688542 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.688814 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4m94s\" (UniqueName: \"kubernetes.io/projected/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b-kube-api-access-4m94s\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.705566 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.719035 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mckr" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.727196 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.734842 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.748052 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905239 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f8e528-9dfd-4572-9ad6-07d8393be1bb-operator-scripts\") pod \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905297 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5efdf59-1cdd-4aef-8105-f44519426d2c-operator-scripts\") pod \"c5efdf59-1cdd-4aef-8105-f44519426d2c\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905329 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s2j6\" (UniqueName: \"kubernetes.io/projected/51f8e528-9dfd-4572-9ad6-07d8393be1bb-kube-api-access-5s2j6\") pod \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\" (UID: \"51f8e528-9dfd-4572-9ad6-07d8393be1bb\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905394 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snfkf\" (UniqueName: \"kubernetes.io/projected/b226f5f5-03ea-4614-bd24-5499b6fe7457-kube-api-access-snfkf\") pod \"b226f5f5-03ea-4614-bd24-5499b6fe7457\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905430 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f077be8-3497-464c-9bfd-4752c6bc08c3-operator-scripts\") pod \"2f077be8-3497-464c-9bfd-4752c6bc08c3\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905446 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddf2w\" (UniqueName: \"kubernetes.io/projected/2f077be8-3497-464c-9bfd-4752c6bc08c3-kube-api-access-ddf2w\") pod \"2f077be8-3497-464c-9bfd-4752c6bc08c3\" (UID: \"2f077be8-3497-464c-9bfd-4752c6bc08c3\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905502 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtk9b\" (UniqueName: \"kubernetes.io/projected/f09ba32d-29da-4907-80b0-4df3433704f8-kube-api-access-vtk9b\") pod \"f09ba32d-29da-4907-80b0-4df3433704f8\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905518 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f09ba32d-29da-4907-80b0-4df3433704f8-operator-scripts\") pod \"f09ba32d-29da-4907-80b0-4df3433704f8\" (UID: \"f09ba32d-29da-4907-80b0-4df3433704f8\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905561 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k5rl\" (UniqueName: \"kubernetes.io/projected/c5efdf59-1cdd-4aef-8105-f44519426d2c-kube-api-access-2k5rl\") pod \"c5efdf59-1cdd-4aef-8105-f44519426d2c\" (UID: \"c5efdf59-1cdd-4aef-8105-f44519426d2c\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.905590 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b226f5f5-03ea-4614-bd24-5499b6fe7457-operator-scripts\") pod \"b226f5f5-03ea-4614-bd24-5499b6fe7457\" (UID: \"b226f5f5-03ea-4614-bd24-5499b6fe7457\") " Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.906218 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b226f5f5-03ea-4614-bd24-5499b6fe7457-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b226f5f5-03ea-4614-bd24-5499b6fe7457" (UID: "b226f5f5-03ea-4614-bd24-5499b6fe7457"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.907326 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f09ba32d-29da-4907-80b0-4df3433704f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f09ba32d-29da-4907-80b0-4df3433704f8" (UID: "f09ba32d-29da-4907-80b0-4df3433704f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.907361 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5efdf59-1cdd-4aef-8105-f44519426d2c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5efdf59-1cdd-4aef-8105-f44519426d2c" (UID: "c5efdf59-1cdd-4aef-8105-f44519426d2c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.907805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f077be8-3497-464c-9bfd-4752c6bc08c3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2f077be8-3497-464c-9bfd-4752c6bc08c3" (UID: "2f077be8-3497-464c-9bfd-4752c6bc08c3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.907941 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51f8e528-9dfd-4572-9ad6-07d8393be1bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "51f8e528-9dfd-4572-9ad6-07d8393be1bb" (UID: "51f8e528-9dfd-4572-9ad6-07d8393be1bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.911716 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b226f5f5-03ea-4614-bd24-5499b6fe7457-kube-api-access-snfkf" (OuterVolumeSpecName: "kube-api-access-snfkf") pod "b226f5f5-03ea-4614-bd24-5499b6fe7457" (UID: "b226f5f5-03ea-4614-bd24-5499b6fe7457"). InnerVolumeSpecName "kube-api-access-snfkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.911793 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51f8e528-9dfd-4572-9ad6-07d8393be1bb-kube-api-access-5s2j6" (OuterVolumeSpecName: "kube-api-access-5s2j6") pod "51f8e528-9dfd-4572-9ad6-07d8393be1bb" (UID: "51f8e528-9dfd-4572-9ad6-07d8393be1bb"). InnerVolumeSpecName "kube-api-access-5s2j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.912031 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5efdf59-1cdd-4aef-8105-f44519426d2c-kube-api-access-2k5rl" (OuterVolumeSpecName: "kube-api-access-2k5rl") pod "c5efdf59-1cdd-4aef-8105-f44519426d2c" (UID: "c5efdf59-1cdd-4aef-8105-f44519426d2c"). InnerVolumeSpecName "kube-api-access-2k5rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.912257 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f09ba32d-29da-4907-80b0-4df3433704f8-kube-api-access-vtk9b" (OuterVolumeSpecName: "kube-api-access-vtk9b") pod "f09ba32d-29da-4907-80b0-4df3433704f8" (UID: "f09ba32d-29da-4907-80b0-4df3433704f8"). InnerVolumeSpecName "kube-api-access-vtk9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.912724 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f077be8-3497-464c-9bfd-4752c6bc08c3-kube-api-access-ddf2w" (OuterVolumeSpecName: "kube-api-access-ddf2w") pod "2f077be8-3497-464c-9bfd-4752c6bc08c3" (UID: "2f077be8-3497-464c-9bfd-4752c6bc08c3"). InnerVolumeSpecName "kube-api-access-ddf2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.987605 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-e1f4-account-create-update-pnscz" event={"ID":"f09ba32d-29da-4907-80b0-4df3433704f8","Type":"ContainerDied","Data":"7070518f1bdbbcdae7c2f5693177bf2b3119eb8775e3a9a07de4db877b303ada"} Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.987654 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7070518f1bdbbcdae7c2f5693177bf2b3119eb8775e3a9a07de4db877b303ada" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.987619 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-e1f4-account-create-update-pnscz" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.989111 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-17f0-account-create-update-8xfqp" event={"ID":"b226f5f5-03ea-4614-bd24-5499b6fe7457","Type":"ContainerDied","Data":"471d2e62258aea5c12e18a7f7b2b058fe0714c0f16d2f240ee3d7f2e4574256d"} Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.989195 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="471d2e62258aea5c12e18a7f7b2b058fe0714c0f16d2f240ee3d7f2e4574256d" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.989142 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-17f0-account-create-update-8xfqp" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.994414 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-98b3-account-create-update-4rg7b" event={"ID":"c5efdf59-1cdd-4aef-8105-f44519426d2c","Type":"ContainerDied","Data":"9c38336d62298d2291ced67070680b1c5951e3a76e9ae53238c5eefa730ed01c"} Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.994463 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c38336d62298d2291ced67070680b1c5951e3a76e9ae53238c5eefa730ed01c" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.994972 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-98b3-account-create-update-4rg7b" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.996105 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-5mckr" event={"ID":"2f077be8-3497-464c-9bfd-4752c6bc08c3","Type":"ContainerDied","Data":"7e2ea59d24fe0530cc81357c2d22c19c29d21e1d08148a171e0a99d4d768f541"} Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.996131 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e2ea59d24fe0530cc81357c2d22c19c29d21e1d08148a171e0a99d4d768f541" Dec 06 10:54:37 crc kubenswrapper[4678]: I1206 10:54:37.996176 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-5mckr" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.001139 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mwspd" event={"ID":"e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b","Type":"ContainerDied","Data":"cdda316e3e748e03aa9ff169e021006f1c02d4d05306e381c9a8286814025ea9"} Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.001169 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdda316e3e748e03aa9ff169e021006f1c02d4d05306e381c9a8286814025ea9" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.001254 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mwspd" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.002619 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-xvjr7" event={"ID":"51f8e528-9dfd-4572-9ad6-07d8393be1bb","Type":"ContainerDied","Data":"43b7d2d156f6b9165d38639ffa0cbaa5a08a1b1c75de6490393670316fdb9958"} Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.002638 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43b7d2d156f6b9165d38639ffa0cbaa5a08a1b1c75de6490393670316fdb9958" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.002767 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-xvjr7" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007431 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s2j6\" (UniqueName: \"kubernetes.io/projected/51f8e528-9dfd-4572-9ad6-07d8393be1bb-kube-api-access-5s2j6\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007475 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snfkf\" (UniqueName: \"kubernetes.io/projected/b226f5f5-03ea-4614-bd24-5499b6fe7457-kube-api-access-snfkf\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007513 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f077be8-3497-464c-9bfd-4752c6bc08c3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007527 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddf2w\" (UniqueName: \"kubernetes.io/projected/2f077be8-3497-464c-9bfd-4752c6bc08c3-kube-api-access-ddf2w\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007539 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtk9b\" (UniqueName: \"kubernetes.io/projected/f09ba32d-29da-4907-80b0-4df3433704f8-kube-api-access-vtk9b\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007550 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f09ba32d-29da-4907-80b0-4df3433704f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007562 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k5rl\" (UniqueName: \"kubernetes.io/projected/c5efdf59-1cdd-4aef-8105-f44519426d2c-kube-api-access-2k5rl\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007574 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b226f5f5-03ea-4614-bd24-5499b6fe7457-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007587 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/51f8e528-9dfd-4572-9ad6-07d8393be1bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:38 crc kubenswrapper[4678]: I1206 10:54:38.007599 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5efdf59-1cdd-4aef-8105-f44519426d2c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:39 crc kubenswrapper[4678]: I1206 10:54:39.016412 4678 generic.go:334] "Generic (PLEG): container finished" podID="995bfee3-8461-4774-aa52-576dc0eacbda" containerID="a67896673bf325c989b177f147cde987c7fc2280ee8ac8bf78a9c554bbb3c2b7" exitCode=0 Dec 06 10:54:39 crc kubenswrapper[4678]: I1206 10:54:39.016509 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"995bfee3-8461-4774-aa52-576dc0eacbda","Type":"ContainerDied","Data":"a67896673bf325c989b177f147cde987c7fc2280ee8ac8bf78a9c554bbb3c2b7"} Dec 06 10:54:40 crc kubenswrapper[4678]: I1206 10:54:40.025924 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"995bfee3-8461-4774-aa52-576dc0eacbda","Type":"ContainerStarted","Data":"649f042958a1f97564fc9864785c913d1ae04d99007549421e0d8f0b8e4378ef"} Dec 06 10:54:40 crc kubenswrapper[4678]: I1206 10:54:40.027266 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:54:40 crc kubenswrapper[4678]: I1206 10:54:40.054033 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.857388688 podStartE2EDuration="1m7.05401254s" podCreationTimestamp="2025-12-06 10:53:33 +0000 UTC" firstStartedPulling="2025-12-06 10:53:35.679983307 +0000 UTC m=+1020.523414746" lastFinishedPulling="2025-12-06 10:54:05.876607149 +0000 UTC m=+1050.720038598" observedRunningTime="2025-12-06 10:54:40.052991599 +0000 UTC m=+1084.896423038" watchObservedRunningTime="2025-12-06 10:54:40.05401254 +0000 UTC m=+1084.897443979" Dec 06 10:54:42 crc kubenswrapper[4678]: I1206 10:54:42.056802 4678 generic.go:334] "Generic (PLEG): container finished" podID="db2561ff-632a-463d-a7af-79d83536812b" containerID="365e9887f82f39fbb4167c9c235cf43f46b2de75ac60f2bbd15d0bfa722bece5" exitCode=0 Dec 06 10:54:42 crc kubenswrapper[4678]: I1206 10:54:42.056893 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tkg52" event={"ID":"db2561ff-632a-463d-a7af-79d83536812b","Type":"ContainerDied","Data":"365e9887f82f39fbb4167c9c235cf43f46b2de75ac60f2bbd15d0bfa722bece5"} Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.380629 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.502733 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-combined-ca-bundle\") pod \"db2561ff-632a-463d-a7af-79d83536812b\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.503001 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-scripts\") pod \"db2561ff-632a-463d-a7af-79d83536812b\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.503105 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-dispersionconf\") pod \"db2561ff-632a-463d-a7af-79d83536812b\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.503198 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-ring-data-devices\") pod \"db2561ff-632a-463d-a7af-79d83536812b\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.503279 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-swiftconf\") pod \"db2561ff-632a-463d-a7af-79d83536812b\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.503525 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkvq5\" (UniqueName: \"kubernetes.io/projected/db2561ff-632a-463d-a7af-79d83536812b-kube-api-access-bkvq5\") pod \"db2561ff-632a-463d-a7af-79d83536812b\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.503678 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2561ff-632a-463d-a7af-79d83536812b-etc-swift\") pod \"db2561ff-632a-463d-a7af-79d83536812b\" (UID: \"db2561ff-632a-463d-a7af-79d83536812b\") " Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.503886 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "db2561ff-632a-463d-a7af-79d83536812b" (UID: "db2561ff-632a-463d-a7af-79d83536812b"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.504347 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db2561ff-632a-463d-a7af-79d83536812b-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "db2561ff-632a-463d-a7af-79d83536812b" (UID: "db2561ff-632a-463d-a7af-79d83536812b"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.509734 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db2561ff-632a-463d-a7af-79d83536812b-kube-api-access-bkvq5" (OuterVolumeSpecName: "kube-api-access-bkvq5") pod "db2561ff-632a-463d-a7af-79d83536812b" (UID: "db2561ff-632a-463d-a7af-79d83536812b"). InnerVolumeSpecName "kube-api-access-bkvq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.512681 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "db2561ff-632a-463d-a7af-79d83536812b" (UID: "db2561ff-632a-463d-a7af-79d83536812b"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.529585 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "db2561ff-632a-463d-a7af-79d83536812b" (UID: "db2561ff-632a-463d-a7af-79d83536812b"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.533033 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "db2561ff-632a-463d-a7af-79d83536812b" (UID: "db2561ff-632a-463d-a7af-79d83536812b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.535911 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-scripts" (OuterVolumeSpecName: "scripts") pod "db2561ff-632a-463d-a7af-79d83536812b" (UID: "db2561ff-632a-463d-a7af-79d83536812b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592274 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-h5wzb"] Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592727 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerName="init" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592755 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerName="init" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592767 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f077be8-3497-464c-9bfd-4752c6bc08c3" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592776 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f077be8-3497-464c-9bfd-4752c6bc08c3" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592792 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b226f5f5-03ea-4614-bd24-5499b6fe7457" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592800 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b226f5f5-03ea-4614-bd24-5499b6fe7457" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592816 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5efdf59-1cdd-4aef-8105-f44519426d2c" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592824 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5efdf59-1cdd-4aef-8105-f44519426d2c" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592833 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09ba32d-29da-4907-80b0-4df3433704f8" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592840 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09ba32d-29da-4907-80b0-4df3433704f8" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592852 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592859 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592872 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51f8e528-9dfd-4572-9ad6-07d8393be1bb" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592880 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="51f8e528-9dfd-4572-9ad6-07d8393be1bb" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592896 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerName="dnsmasq-dns" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592904 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerName="dnsmasq-dns" Dec 06 10:54:43 crc kubenswrapper[4678]: E1206 10:54:43.592915 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db2561ff-632a-463d-a7af-79d83536812b" containerName="swift-ring-rebalance" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.592922 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="db2561ff-632a-463d-a7af-79d83536812b" containerName="swift-ring-rebalance" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593112 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac16ad4-74b1-4968-bd23-10c814ed087d" containerName="dnsmasq-dns" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593133 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593144 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5efdf59-1cdd-4aef-8105-f44519426d2c" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593154 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b226f5f5-03ea-4614-bd24-5499b6fe7457" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593167 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="51f8e528-9dfd-4572-9ad6-07d8393be1bb" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593179 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="db2561ff-632a-463d-a7af-79d83536812b" containerName="swift-ring-rebalance" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593190 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f077be8-3497-464c-9bfd-4752c6bc08c3" containerName="mariadb-database-create" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593201 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09ba32d-29da-4907-80b0-4df3433704f8" containerName="mariadb-account-create-update" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.593861 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.596637 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.597628 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xrkxl" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.606641 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.606889 4678 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.606988 4678 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/db2561ff-632a-463d-a7af-79d83536812b-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.607080 4678 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.607263 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkvq5\" (UniqueName: \"kubernetes.io/projected/db2561ff-632a-463d-a7af-79d83536812b-kube-api-access-bkvq5\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.607345 4678 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/db2561ff-632a-463d-a7af-79d83536812b-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.607429 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/db2561ff-632a-463d-a7af-79d83536812b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.620943 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h5wzb"] Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.708964 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-db-sync-config-data\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.709266 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-combined-ca-bundle\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.709427 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-config-data\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.710599 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66mn6\" (UniqueName: \"kubernetes.io/projected/9024ed1c-7251-4e71-aa5e-48354a9813b7-kube-api-access-66mn6\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.811872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-db-sync-config-data\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.812219 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-combined-ca-bundle\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.812565 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-config-data\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.812740 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66mn6\" (UniqueName: \"kubernetes.io/projected/9024ed1c-7251-4e71-aa5e-48354a9813b7-kube-api-access-66mn6\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.817042 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-db-sync-config-data\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.817778 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-combined-ca-bundle\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.820793 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-config-data\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.844603 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66mn6\" (UniqueName: \"kubernetes.io/projected/9024ed1c-7251-4e71-aa5e-48354a9813b7-kube-api-access-66mn6\") pod \"glance-db-sync-h5wzb\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:43 crc kubenswrapper[4678]: I1206 10:54:43.911925 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h5wzb" Dec 06 10:54:44 crc kubenswrapper[4678]: I1206 10:54:44.075105 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-tkg52" event={"ID":"db2561ff-632a-463d-a7af-79d83536812b","Type":"ContainerDied","Data":"220d7d9e72986693b1b6bd63085e832e17654e1c1f86b1f22b155321d1997150"} Dec 06 10:54:44 crc kubenswrapper[4678]: I1206 10:54:44.075457 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="220d7d9e72986693b1b6bd63085e832e17654e1c1f86b1f22b155321d1997150" Dec 06 10:54:44 crc kubenswrapper[4678]: I1206 10:54:44.075226 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-tkg52" Dec 06 10:54:44 crc kubenswrapper[4678]: I1206 10:54:44.497967 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h5wzb"] Dec 06 10:54:45 crc kubenswrapper[4678]: I1206 10:54:45.086276 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h5wzb" event={"ID":"9024ed1c-7251-4e71-aa5e-48354a9813b7","Type":"ContainerStarted","Data":"3bf9a73bd056e69cc084fb1f446e541ebf042bb239a32d538489fb943b833695"} Dec 06 10:54:45 crc kubenswrapper[4678]: I1206 10:54:45.632635 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.143918 4678 generic.go:334] "Generic (PLEG): container finished" podID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerID="f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b" exitCode=0 Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.144007 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52efc5a4-585a-4b9e-93f9-f40527e9a6c9","Type":"ContainerDied","Data":"f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b"} Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.251340 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kdb5q" podUID="c2658168-1355-4595-bf23-570a68c90da1" containerName="ovn-controller" probeResult="failure" output=< Dec 06 10:54:50 crc kubenswrapper[4678]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 10:54:50 crc kubenswrapper[4678]: > Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.260270 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.269962 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zcg6x" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.499043 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-kdb5q-config-sf29h"] Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.500394 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.502797 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.508864 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kdb5q-config-sf29h"] Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.641794 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.642149 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-additional-scripts\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.642190 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xpft\" (UniqueName: \"kubernetes.io/projected/4eb288a8-1dbc-409b-80ee-29c817db8d4f-kube-api-access-4xpft\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.642237 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run-ovn\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.642261 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-log-ovn\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.642310 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-scripts\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.743237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.743304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-additional-scripts\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.743337 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xpft\" (UniqueName: \"kubernetes.io/projected/4eb288a8-1dbc-409b-80ee-29c817db8d4f-kube-api-access-4xpft\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.743373 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run-ovn\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.743392 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-log-ovn\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.743441 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-scripts\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.743991 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.744057 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-log-ovn\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.744134 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run-ovn\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.744575 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-additional-scripts\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.745510 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-scripts\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.775546 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xpft\" (UniqueName: \"kubernetes.io/projected/4eb288a8-1dbc-409b-80ee-29c817db8d4f-kube-api-access-4xpft\") pod \"ovn-controller-kdb5q-config-sf29h\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:50 crc kubenswrapper[4678]: I1206 10:54:50.833600 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:53 crc kubenswrapper[4678]: I1206 10:54:53.303591 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:53 crc kubenswrapper[4678]: I1206 10:54:53.312790 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/a7f4ca97-1480-4327-8025-9b18a8a7073b-etc-swift\") pod \"swift-storage-0\" (UID: \"a7f4ca97-1480-4327-8025-9b18a8a7073b\") " pod="openstack/swift-storage-0" Dec 06 10:54:53 crc kubenswrapper[4678]: I1206 10:54:53.522845 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 06 10:54:55 crc kubenswrapper[4678]: I1206 10:54:55.069783 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 10:54:55 crc kubenswrapper[4678]: I1206 10:54:55.321728 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-kdb5q" podUID="c2658168-1355-4595-bf23-570a68c90da1" containerName="ovn-controller" probeResult="failure" output=< Dec 06 10:54:55 crc kubenswrapper[4678]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 06 10:54:55 crc kubenswrapper[4678]: > Dec 06 10:54:56 crc kubenswrapper[4678]: I1206 10:54:56.574943 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-kdb5q-config-sf29h"] Dec 06 10:54:56 crc kubenswrapper[4678]: I1206 10:54:56.833979 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.216637 4678 generic.go:334] "Generic (PLEG): container finished" podID="4eb288a8-1dbc-409b-80ee-29c817db8d4f" containerID="130cf8c56956d6e41ce3133a71346ec38be43b34389acfd82e645c06b8e6e1d0" exitCode=0 Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.217351 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kdb5q-config-sf29h" event={"ID":"4eb288a8-1dbc-409b-80ee-29c817db8d4f","Type":"ContainerDied","Data":"130cf8c56956d6e41ce3133a71346ec38be43b34389acfd82e645c06b8e6e1d0"} Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.217379 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kdb5q-config-sf29h" event={"ID":"4eb288a8-1dbc-409b-80ee-29c817db8d4f","Type":"ContainerStarted","Data":"f53aaf9dbc8fc990c31f4580ba20d661c1450f705a3ac7fad803826759621fe1"} Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.218539 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"9a3a8c87c137f9c9b2fb58d15c70571c60bb0ec45f9f4f150f3865e3ee8d1fc8"} Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.221662 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52efc5a4-585a-4b9e-93f9-f40527e9a6c9","Type":"ContainerStarted","Data":"02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4"} Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.222610 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.223760 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h5wzb" event={"ID":"9024ed1c-7251-4e71-aa5e-48354a9813b7","Type":"ContainerStarted","Data":"0ccf83f93468cdad88fec3e6b3dc40dc30e3dab86044fbf29afd783baf1842f2"} Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.266256 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371952.588535 podStartE2EDuration="1m24.2662398s" podCreationTimestamp="2025-12-06 10:53:33 +0000 UTC" firstStartedPulling="2025-12-06 10:53:35.371033659 +0000 UTC m=+1020.214465098" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:54:57.25968322 +0000 UTC m=+1102.103114649" watchObservedRunningTime="2025-12-06 10:54:57.2662398 +0000 UTC m=+1102.109671239" Dec 06 10:54:57 crc kubenswrapper[4678]: I1206 10:54:57.280434 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-h5wzb" podStartSLOduration=2.5273979989999997 podStartE2EDuration="14.280418745s" podCreationTimestamp="2025-12-06 10:54:43 +0000 UTC" firstStartedPulling="2025-12-06 10:54:44.512359732 +0000 UTC m=+1089.355791171" lastFinishedPulling="2025-12-06 10:54:56.265380478 +0000 UTC m=+1101.108811917" observedRunningTime="2025-12-06 10:54:57.276062942 +0000 UTC m=+1102.119494371" watchObservedRunningTime="2025-12-06 10:54:57.280418745 +0000 UTC m=+1102.123850184" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.234904 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"2a6fdbb869ece1b923b0bd5795e093c472fbe5825b21f4a29b58f54a3e612d98"} Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.235191 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"3c815b5b85426f5de6a4c9fad05858f123fa5205433ae44649e2ac12da2a3d04"} Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.595510 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.714857 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-additional-scripts\") pod \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.714960 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-scripts\") pod \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.715000 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-log-ovn\") pod \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.715035 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run\") pod \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.715066 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run-ovn\") pod \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.715172 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xpft\" (UniqueName: \"kubernetes.io/projected/4eb288a8-1dbc-409b-80ee-29c817db8d4f-kube-api-access-4xpft\") pod \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\" (UID: \"4eb288a8-1dbc-409b-80ee-29c817db8d4f\") " Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.716531 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4eb288a8-1dbc-409b-80ee-29c817db8d4f" (UID: "4eb288a8-1dbc-409b-80ee-29c817db8d4f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.716582 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4eb288a8-1dbc-409b-80ee-29c817db8d4f" (UID: "4eb288a8-1dbc-409b-80ee-29c817db8d4f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.716798 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4eb288a8-1dbc-409b-80ee-29c817db8d4f" (UID: "4eb288a8-1dbc-409b-80ee-29c817db8d4f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.717284 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-scripts" (OuterVolumeSpecName: "scripts") pod "4eb288a8-1dbc-409b-80ee-29c817db8d4f" (UID: "4eb288a8-1dbc-409b-80ee-29c817db8d4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.717357 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run" (OuterVolumeSpecName: "var-run") pod "4eb288a8-1dbc-409b-80ee-29c817db8d4f" (UID: "4eb288a8-1dbc-409b-80ee-29c817db8d4f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.722648 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eb288a8-1dbc-409b-80ee-29c817db8d4f-kube-api-access-4xpft" (OuterVolumeSpecName: "kube-api-access-4xpft") pod "4eb288a8-1dbc-409b-80ee-29c817db8d4f" (UID: "4eb288a8-1dbc-409b-80ee-29c817db8d4f"). InnerVolumeSpecName "kube-api-access-4xpft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.817087 4678 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.817134 4678 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.817149 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xpft\" (UniqueName: \"kubernetes.io/projected/4eb288a8-1dbc-409b-80ee-29c817db8d4f-kube-api-access-4xpft\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.817162 4678 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.817174 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eb288a8-1dbc-409b-80ee-29c817db8d4f-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:58 crc kubenswrapper[4678]: I1206 10:54:58.817185 4678 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4eb288a8-1dbc-409b-80ee-29c817db8d4f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.244706 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-kdb5q-config-sf29h" event={"ID":"4eb288a8-1dbc-409b-80ee-29c817db8d4f","Type":"ContainerDied","Data":"f53aaf9dbc8fc990c31f4580ba20d661c1450f705a3ac7fad803826759621fe1"} Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.244757 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f53aaf9dbc8fc990c31f4580ba20d661c1450f705a3ac7fad803826759621fe1" Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.244937 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-kdb5q-config-sf29h" Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.247250 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"89fae94615163c2254405d826b179ae26e3a1f437d7f82c3bb659b11ba3e7b57"} Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.247297 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"2c85672cf4d820c47cc42432c0220c232ce2f06c1bcced074618fa060d5b6b20"} Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.505936 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.505996 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.705210 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-kdb5q-config-sf29h"] Dec 06 10:54:59 crc kubenswrapper[4678]: I1206 10:54:59.714381 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-kdb5q-config-sf29h"] Dec 06 10:55:00 crc kubenswrapper[4678]: I1206 10:55:00.306780 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-kdb5q" Dec 06 10:55:01 crc kubenswrapper[4678]: I1206 10:55:01.273503 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"cca098bda240952c640efe3213190f22b9507bb94a95b0f9d29c8ef8ee685f9c"} Dec 06 10:55:01 crc kubenswrapper[4678]: I1206 10:55:01.273779 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"2fc39534047bd736cc9b9f9eb6afbe1c390e3df2b4058eec470353f7078b655b"} Dec 06 10:55:01 crc kubenswrapper[4678]: I1206 10:55:01.273790 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"aef9252321958a1ffb07b5c717f8e11a624d0f0ab5642745bc4d5b4d723086b2"} Dec 06 10:55:01 crc kubenswrapper[4678]: I1206 10:55:01.273800 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"b3b7fdbc72b87800440d1f34df1df689e093d04dce45a1821a911645fa8735dd"} Dec 06 10:55:01 crc kubenswrapper[4678]: I1206 10:55:01.486905 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eb288a8-1dbc-409b-80ee-29c817db8d4f" path="/var/lib/kubelet/pods/4eb288a8-1dbc-409b-80ee-29c817db8d4f/volumes" Dec 06 10:55:02 crc kubenswrapper[4678]: I1206 10:55:02.287413 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"41a09a6c5076b3de1d00533cef519c8e4b8ea29a25dff414a4480ebc2fc13c53"} Dec 06 10:55:03 crc kubenswrapper[4678]: I1206 10:55:03.313162 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"73f47c24d5d2ef66d38d04a319d8ec683ce3bb08d51b07c43fa0cbe66636a3fd"} Dec 06 10:55:03 crc kubenswrapper[4678]: I1206 10:55:03.313218 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"fdf7b0c667823bdff91445c7bb2334815b344fd50fe6e4af3719ec6acc5b8c82"} Dec 06 10:55:03 crc kubenswrapper[4678]: I1206 10:55:03.313236 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"dd086ed10faf6a9d65c0cb41a1f86913d30389ea3953213ace02e470ae35ef3a"} Dec 06 10:55:03 crc kubenswrapper[4678]: I1206 10:55:03.313248 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"1ed88e835c9cc538616b305b4fca918d0bb3c0d8ffc1ef6d1f3e1c8bab1e28bc"} Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.328747 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"fe88e9c468c25b5db14c8a8e8cd8cafa0c4327c64346f1f578754bab214dddd4"} Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.329047 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"a7f4ca97-1480-4327-8025-9b18a8a7073b","Type":"ContainerStarted","Data":"752fa930217c98d0a733c659769eeb3ae67f63f91b9329a9febd73b656926d1f"} Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.648192 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=39.385732927 podStartE2EDuration="44.648175576s" podCreationTimestamp="2025-12-06 10:54:20 +0000 UTC" firstStartedPulling="2025-12-06 10:54:56.841865235 +0000 UTC m=+1101.685296674" lastFinishedPulling="2025-12-06 10:55:02.104307884 +0000 UTC m=+1106.947739323" observedRunningTime="2025-12-06 10:55:04.367117485 +0000 UTC m=+1109.210548924" watchObservedRunningTime="2025-12-06 10:55:04.648175576 +0000 UTC m=+1109.491607015" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.650661 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-pb8tc"] Dec 06 10:55:04 crc kubenswrapper[4678]: E1206 10:55:04.650962 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eb288a8-1dbc-409b-80ee-29c817db8d4f" containerName="ovn-config" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.650978 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eb288a8-1dbc-409b-80ee-29c817db8d4f" containerName="ovn-config" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.651207 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eb288a8-1dbc-409b-80ee-29c817db8d4f" containerName="ovn-config" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.652059 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.654906 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.669021 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-pb8tc"] Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.813740 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/0746a366-e454-4add-b80c-020dc0a1d07a-kube-api-access-fvnsz\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.813809 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.813891 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-config\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.813979 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.814115 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.814276 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.916398 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.916546 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.916643 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/0746a366-e454-4add-b80c-020dc0a1d07a-kube-api-access-fvnsz\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.916690 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.916737 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-config\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.916767 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.917319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.917427 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.917782 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.917782 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.918084 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-config\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.951711 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/0746a366-e454-4add-b80c-020dc0a1d07a-kube-api-access-fvnsz\") pod \"dnsmasq-dns-5c79d794d7-pb8tc\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:04 crc kubenswrapper[4678]: I1206 10:55:04.972601 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:05 crc kubenswrapper[4678]: I1206 10:55:05.462754 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-pb8tc"] Dec 06 10:55:06 crc kubenswrapper[4678]: I1206 10:55:06.346354 4678 generic.go:334] "Generic (PLEG): container finished" podID="9024ed1c-7251-4e71-aa5e-48354a9813b7" containerID="0ccf83f93468cdad88fec3e6b3dc40dc30e3dab86044fbf29afd783baf1842f2" exitCode=0 Dec 06 10:55:06 crc kubenswrapper[4678]: I1206 10:55:06.346453 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h5wzb" event={"ID":"9024ed1c-7251-4e71-aa5e-48354a9813b7","Type":"ContainerDied","Data":"0ccf83f93468cdad88fec3e6b3dc40dc30e3dab86044fbf29afd783baf1842f2"} Dec 06 10:55:06 crc kubenswrapper[4678]: I1206 10:55:06.349094 4678 generic.go:334] "Generic (PLEG): container finished" podID="0746a366-e454-4add-b80c-020dc0a1d07a" containerID="e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1" exitCode=0 Dec 06 10:55:06 crc kubenswrapper[4678]: I1206 10:55:06.349135 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" event={"ID":"0746a366-e454-4add-b80c-020dc0a1d07a","Type":"ContainerDied","Data":"e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1"} Dec 06 10:55:06 crc kubenswrapper[4678]: I1206 10:55:06.349164 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" event={"ID":"0746a366-e454-4add-b80c-020dc0a1d07a","Type":"ContainerStarted","Data":"d1c04db193caf2b519f8d1fbe3ed72436d67927607d66817e66c2604e2040a08"} Dec 06 10:55:07 crc kubenswrapper[4678]: I1206 10:55:07.364200 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" event={"ID":"0746a366-e454-4add-b80c-020dc0a1d07a","Type":"ContainerStarted","Data":"2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd"} Dec 06 10:55:07 crc kubenswrapper[4678]: I1206 10:55:07.401680 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" podStartSLOduration=3.401477731 podStartE2EDuration="3.401477731s" podCreationTimestamp="2025-12-06 10:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:55:07.399588786 +0000 UTC m=+1112.243020235" watchObservedRunningTime="2025-12-06 10:55:07.401477731 +0000 UTC m=+1112.244909180" Dec 06 10:55:07 crc kubenswrapper[4678]: I1206 10:55:07.918007 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h5wzb" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.080352 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-config-data\") pod \"9024ed1c-7251-4e71-aa5e-48354a9813b7\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.080561 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66mn6\" (UniqueName: \"kubernetes.io/projected/9024ed1c-7251-4e71-aa5e-48354a9813b7-kube-api-access-66mn6\") pod \"9024ed1c-7251-4e71-aa5e-48354a9813b7\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.080660 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-db-sync-config-data\") pod \"9024ed1c-7251-4e71-aa5e-48354a9813b7\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.080705 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-combined-ca-bundle\") pod \"9024ed1c-7251-4e71-aa5e-48354a9813b7\" (UID: \"9024ed1c-7251-4e71-aa5e-48354a9813b7\") " Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.094999 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9024ed1c-7251-4e71-aa5e-48354a9813b7" (UID: "9024ed1c-7251-4e71-aa5e-48354a9813b7"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.095029 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9024ed1c-7251-4e71-aa5e-48354a9813b7-kube-api-access-66mn6" (OuterVolumeSpecName: "kube-api-access-66mn6") pod "9024ed1c-7251-4e71-aa5e-48354a9813b7" (UID: "9024ed1c-7251-4e71-aa5e-48354a9813b7"). InnerVolumeSpecName "kube-api-access-66mn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.102885 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9024ed1c-7251-4e71-aa5e-48354a9813b7" (UID: "9024ed1c-7251-4e71-aa5e-48354a9813b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.127156 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-config-data" (OuterVolumeSpecName: "config-data") pod "9024ed1c-7251-4e71-aa5e-48354a9813b7" (UID: "9024ed1c-7251-4e71-aa5e-48354a9813b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.183592 4678 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.183634 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.183648 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9024ed1c-7251-4e71-aa5e-48354a9813b7-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.183659 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66mn6\" (UniqueName: \"kubernetes.io/projected/9024ed1c-7251-4e71-aa5e-48354a9813b7-kube-api-access-66mn6\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.393164 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h5wzb" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.393227 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h5wzb" event={"ID":"9024ed1c-7251-4e71-aa5e-48354a9813b7","Type":"ContainerDied","Data":"3bf9a73bd056e69cc084fb1f446e541ebf042bb239a32d538489fb943b833695"} Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.393260 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bf9a73bd056e69cc084fb1f446e541ebf042bb239a32d538489fb943b833695" Dec 06 10:55:08 crc kubenswrapper[4678]: I1206 10:55:08.394640 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.028615 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-pb8tc"] Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.052370 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-x6q8z"] Dec 06 10:55:09 crc kubenswrapper[4678]: E1206 10:55:09.052935 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9024ed1c-7251-4e71-aa5e-48354a9813b7" containerName="glance-db-sync" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.053002 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9024ed1c-7251-4e71-aa5e-48354a9813b7" containerName="glance-db-sync" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.053245 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9024ed1c-7251-4e71-aa5e-48354a9813b7" containerName="glance-db-sync" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.054205 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.076882 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-x6q8z"] Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.199419 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.199499 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.199553 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.199720 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.199776 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcm7h\" (UniqueName: \"kubernetes.io/projected/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-kube-api-access-lcm7h\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.199843 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-config\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.301348 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.301432 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.301465 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.302354 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.302842 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.302904 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.303663 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.303691 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcm7h\" (UniqueName: \"kubernetes.io/projected/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-kube-api-access-lcm7h\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.304219 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.304633 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-config\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.305914 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-config\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.340352 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcm7h\" (UniqueName: \"kubernetes.io/projected/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-kube-api-access-lcm7h\") pod \"dnsmasq-dns-5f59b8f679-x6q8z\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.371558 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:09 crc kubenswrapper[4678]: I1206 10:55:09.835463 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-x6q8z"] Dec 06 10:55:09 crc kubenswrapper[4678]: W1206 10:55:09.839279 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b1b5d12_ab97_49f9_b355_8da2966ff0bc.slice/crio-8453b948385999c8d39ffedbfcc90db5084ffc2f1a1aa879c133f32e8865dddf WatchSource:0}: Error finding container 8453b948385999c8d39ffedbfcc90db5084ffc2f1a1aa879c133f32e8865dddf: Status 404 returned error can't find the container with id 8453b948385999c8d39ffedbfcc90db5084ffc2f1a1aa879c133f32e8865dddf Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.414111 4678 generic.go:334] "Generic (PLEG): container finished" podID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerID="a0deb11492465300ef7937c8d1f139355d7646e2849951ce842e5cce8ffdcd91" exitCode=0 Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.414637 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" podUID="0746a366-e454-4add-b80c-020dc0a1d07a" containerName="dnsmasq-dns" containerID="cri-o://2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd" gracePeriod=10 Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.414250 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" event={"ID":"2b1b5d12-ab97-49f9-b355-8da2966ff0bc","Type":"ContainerDied","Data":"a0deb11492465300ef7937c8d1f139355d7646e2849951ce842e5cce8ffdcd91"} Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.415080 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" event={"ID":"2b1b5d12-ab97-49f9-b355-8da2966ff0bc","Type":"ContainerStarted","Data":"8453b948385999c8d39ffedbfcc90db5084ffc2f1a1aa879c133f32e8865dddf"} Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.829918 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.931592 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-nb\") pod \"0746a366-e454-4add-b80c-020dc0a1d07a\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.931721 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-sb\") pod \"0746a366-e454-4add-b80c-020dc0a1d07a\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.931757 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-config\") pod \"0746a366-e454-4add-b80c-020dc0a1d07a\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.931832 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-svc\") pod \"0746a366-e454-4add-b80c-020dc0a1d07a\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.931893 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-swift-storage-0\") pod \"0746a366-e454-4add-b80c-020dc0a1d07a\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.931927 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/0746a366-e454-4add-b80c-020dc0a1d07a-kube-api-access-fvnsz\") pod \"0746a366-e454-4add-b80c-020dc0a1d07a\" (UID: \"0746a366-e454-4add-b80c-020dc0a1d07a\") " Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.937397 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0746a366-e454-4add-b80c-020dc0a1d07a-kube-api-access-fvnsz" (OuterVolumeSpecName: "kube-api-access-fvnsz") pod "0746a366-e454-4add-b80c-020dc0a1d07a" (UID: "0746a366-e454-4add-b80c-020dc0a1d07a"). InnerVolumeSpecName "kube-api-access-fvnsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:10 crc kubenswrapper[4678]: I1206 10:55:10.994343 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0746a366-e454-4add-b80c-020dc0a1d07a" (UID: "0746a366-e454-4add-b80c-020dc0a1d07a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.002791 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0746a366-e454-4add-b80c-020dc0a1d07a" (UID: "0746a366-e454-4add-b80c-020dc0a1d07a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.010271 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-config" (OuterVolumeSpecName: "config") pod "0746a366-e454-4add-b80c-020dc0a1d07a" (UID: "0746a366-e454-4add-b80c-020dc0a1d07a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.012982 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0746a366-e454-4add-b80c-020dc0a1d07a" (UID: "0746a366-e454-4add-b80c-020dc0a1d07a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.014003 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0746a366-e454-4add-b80c-020dc0a1d07a" (UID: "0746a366-e454-4add-b80c-020dc0a1d07a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.033660 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.033705 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.033716 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.033728 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.033737 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0746a366-e454-4add-b80c-020dc0a1d07a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.033746 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fvnsz\" (UniqueName: \"kubernetes.io/projected/0746a366-e454-4add-b80c-020dc0a1d07a-kube-api-access-fvnsz\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.423894 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" event={"ID":"2b1b5d12-ab97-49f9-b355-8da2966ff0bc","Type":"ContainerStarted","Data":"7171e508824f9159b9c4a5754f92a0faa2784942ffcfb0f12f71effb41660b8b"} Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.424330 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.426107 4678 generic.go:334] "Generic (PLEG): container finished" podID="0746a366-e454-4add-b80c-020dc0a1d07a" containerID="2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd" exitCode=0 Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.426144 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" event={"ID":"0746a366-e454-4add-b80c-020dc0a1d07a","Type":"ContainerDied","Data":"2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd"} Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.426171 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" event={"ID":"0746a366-e454-4add-b80c-020dc0a1d07a","Type":"ContainerDied","Data":"d1c04db193caf2b519f8d1fbe3ed72436d67927607d66817e66c2604e2040a08"} Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.426210 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-pb8tc" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.426202 4678 scope.go:117] "RemoveContainer" containerID="2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.460480 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" podStartSLOduration=2.46046131 podStartE2EDuration="2.46046131s" podCreationTimestamp="2025-12-06 10:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:55:11.458178593 +0000 UTC m=+1116.301610042" watchObservedRunningTime="2025-12-06 10:55:11.46046131 +0000 UTC m=+1116.303892749" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.469579 4678 scope.go:117] "RemoveContainer" containerID="e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.488705 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-pb8tc"] Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.492306 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-pb8tc"] Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.499696 4678 scope.go:117] "RemoveContainer" containerID="2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd" Dec 06 10:55:11 crc kubenswrapper[4678]: E1206 10:55:11.500261 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd\": container with ID starting with 2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd not found: ID does not exist" containerID="2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.500298 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd"} err="failed to get container status \"2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd\": rpc error: code = NotFound desc = could not find container \"2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd\": container with ID starting with 2969512fb0e7393e55a4ecf7890ed2c2b895a6af70f0e2a6750ba286d77c8efd not found: ID does not exist" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.500324 4678 scope.go:117] "RemoveContainer" containerID="e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1" Dec 06 10:55:11 crc kubenswrapper[4678]: E1206 10:55:11.500644 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1\": container with ID starting with e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1 not found: ID does not exist" containerID="e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1" Dec 06 10:55:11 crc kubenswrapper[4678]: I1206 10:55:11.500666 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1"} err="failed to get container status \"e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1\": rpc error: code = NotFound desc = could not find container \"e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1\": container with ID starting with e74d6b23c0c19d4aaaf9a5c9d4b0955587433a05d94701671c95c666514214e1 not found: ID does not exist" Dec 06 10:55:13 crc kubenswrapper[4678]: I1206 10:55:13.487266 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0746a366-e454-4add-b80c-020dc0a1d07a" path="/var/lib/kubelet/pods/0746a366-e454-4add-b80c-020dc0a1d07a/volumes" Dec 06 10:55:14 crc kubenswrapper[4678]: I1206 10:55:14.653014 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.079141 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qjg97"] Dec 06 10:55:15 crc kubenswrapper[4678]: E1206 10:55:15.079774 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0746a366-e454-4add-b80c-020dc0a1d07a" containerName="dnsmasq-dns" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.079791 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0746a366-e454-4add-b80c-020dc0a1d07a" containerName="dnsmasq-dns" Dec 06 10:55:15 crc kubenswrapper[4678]: E1206 10:55:15.079825 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0746a366-e454-4add-b80c-020dc0a1d07a" containerName="init" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.079831 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0746a366-e454-4add-b80c-020dc0a1d07a" containerName="init" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.079982 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="0746a366-e454-4add-b80c-020dc0a1d07a" containerName="dnsmasq-dns" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.080558 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.091274 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-0275-account-create-update-sxr85"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.093303 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.099361 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.125251 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0275-account-create-update-sxr85"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.146171 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qjg97"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.180162 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-m54tp"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.181852 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.212595 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-operator-scripts\") pod \"cinder-db-create-qjg97\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.212909 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkvdr\" (UniqueName: \"kubernetes.io/projected/daddc760-3153-40b0-8e44-01e32a9d909f-kube-api-access-xkvdr\") pod \"cinder-0275-account-create-update-sxr85\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.213046 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsqsk\" (UniqueName: \"kubernetes.io/projected/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-kube-api-access-fsqsk\") pod \"cinder-db-create-qjg97\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.213190 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/daddc760-3153-40b0-8e44-01e32a9d909f-operator-scripts\") pod \"cinder-0275-account-create-update-sxr85\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.215405 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-055f-account-create-update-wmjpx"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.216396 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.219893 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.260553 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-055f-account-create-update-wmjpx"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.279254 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m54tp"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314598 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pglbz\" (UniqueName: \"kubernetes.io/projected/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-kube-api-access-pglbz\") pod \"barbican-db-create-m54tp\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314650 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h488v\" (UniqueName: \"kubernetes.io/projected/8a074406-0662-4e37-8cb5-4f187d687cb8-kube-api-access-h488v\") pod \"barbican-055f-account-create-update-wmjpx\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314689 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/daddc760-3153-40b0-8e44-01e32a9d909f-operator-scripts\") pod \"cinder-0275-account-create-update-sxr85\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314732 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a074406-0662-4e37-8cb5-4f187d687cb8-operator-scripts\") pod \"barbican-055f-account-create-update-wmjpx\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314781 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-operator-scripts\") pod \"cinder-db-create-qjg97\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314812 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkvdr\" (UniqueName: \"kubernetes.io/projected/daddc760-3153-40b0-8e44-01e32a9d909f-kube-api-access-xkvdr\") pod \"cinder-0275-account-create-update-sxr85\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314839 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-operator-scripts\") pod \"barbican-db-create-m54tp\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.314870 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsqsk\" (UniqueName: \"kubernetes.io/projected/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-kube-api-access-fsqsk\") pod \"cinder-db-create-qjg97\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.315272 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/daddc760-3153-40b0-8e44-01e32a9d909f-operator-scripts\") pod \"cinder-0275-account-create-update-sxr85\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.315807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-operator-scripts\") pod \"cinder-db-create-qjg97\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.339593 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsqsk\" (UniqueName: \"kubernetes.io/projected/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-kube-api-access-fsqsk\") pod \"cinder-db-create-qjg97\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.374687 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkvdr\" (UniqueName: \"kubernetes.io/projected/daddc760-3153-40b0-8e44-01e32a9d909f-kube-api-access-xkvdr\") pod \"cinder-0275-account-create-update-sxr85\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.395990 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.413258 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.416188 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a074406-0662-4e37-8cb5-4f187d687cb8-operator-scripts\") pod \"barbican-055f-account-create-update-wmjpx\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.416259 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-operator-scripts\") pod \"barbican-db-create-m54tp\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.416344 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pglbz\" (UniqueName: \"kubernetes.io/projected/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-kube-api-access-pglbz\") pod \"barbican-db-create-m54tp\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.416375 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h488v\" (UniqueName: \"kubernetes.io/projected/8a074406-0662-4e37-8cb5-4f187d687cb8-kube-api-access-h488v\") pod \"barbican-055f-account-create-update-wmjpx\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.416977 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a074406-0662-4e37-8cb5-4f187d687cb8-operator-scripts\") pod \"barbican-055f-account-create-update-wmjpx\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.417395 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-operator-scripts\") pod \"barbican-db-create-m54tp\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.462939 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-mzqdh"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.466196 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.480237 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z5rpk" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.502659 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pglbz\" (UniqueName: \"kubernetes.io/projected/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-kube-api-access-pglbz\") pod \"barbican-db-create-m54tp\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.504613 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.505117 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.505403 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.519329 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h488v\" (UniqueName: \"kubernetes.io/projected/8a074406-0662-4e37-8cb5-4f187d687cb8-kube-api-access-h488v\") pod \"barbican-055f-account-create-update-wmjpx\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.521588 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.526841 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mzqdh"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.544707 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-x9jhf"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.554933 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.582958 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.632629 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-config-data\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.632720 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d28bdf3-353f-444d-b401-ff06978869c9-operator-scripts\") pod \"neutron-db-create-x9jhf\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.632780 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsqp4\" (UniqueName: \"kubernetes.io/projected/26bbb0e2-894e-4476-b5c8-4af78dedafd8-kube-api-access-dsqp4\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.632824 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkrg2\" (UniqueName: \"kubernetes.io/projected/7d28bdf3-353f-444d-b401-ff06978869c9-kube-api-access-kkrg2\") pod \"neutron-db-create-x9jhf\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.632849 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-combined-ca-bundle\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.705152 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x9jhf"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.771741 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-config-data\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.771799 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d28bdf3-353f-444d-b401-ff06978869c9-operator-scripts\") pod \"neutron-db-create-x9jhf\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.771834 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsqp4\" (UniqueName: \"kubernetes.io/projected/26bbb0e2-894e-4476-b5c8-4af78dedafd8-kube-api-access-dsqp4\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.771852 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkrg2\" (UniqueName: \"kubernetes.io/projected/7d28bdf3-353f-444d-b401-ff06978869c9-kube-api-access-kkrg2\") pod \"neutron-db-create-x9jhf\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.771873 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-combined-ca-bundle\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.773248 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d28bdf3-353f-444d-b401-ff06978869c9-operator-scripts\") pod \"neutron-db-create-x9jhf\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.783520 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f054-account-create-update-f98s6"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.784634 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.786781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-combined-ca-bundle\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.792969 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-config-data\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.801958 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.802195 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f054-account-create-update-f98s6"] Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.811523 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkrg2\" (UniqueName: \"kubernetes.io/projected/7d28bdf3-353f-444d-b401-ff06978869c9-kube-api-access-kkrg2\") pod \"neutron-db-create-x9jhf\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.837213 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsqp4\" (UniqueName: \"kubernetes.io/projected/26bbb0e2-894e-4476-b5c8-4af78dedafd8-kube-api-access-dsqp4\") pod \"keystone-db-sync-mzqdh\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.878156 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97386078-2478-4899-be49-4f8bb3019081-operator-scripts\") pod \"neutron-f054-account-create-update-f98s6\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.885976 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l5zg\" (UniqueName: \"kubernetes.io/projected/97386078-2478-4899-be49-4f8bb3019081-kube-api-access-4l5zg\") pod \"neutron-f054-account-create-update-f98s6\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.902560 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.984976 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.987621 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97386078-2478-4899-be49-4f8bb3019081-operator-scripts\") pod \"neutron-f054-account-create-update-f98s6\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.987679 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l5zg\" (UniqueName: \"kubernetes.io/projected/97386078-2478-4899-be49-4f8bb3019081-kube-api-access-4l5zg\") pod \"neutron-f054-account-create-update-f98s6\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:15 crc kubenswrapper[4678]: I1206 10:55:15.988873 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97386078-2478-4899-be49-4f8bb3019081-operator-scripts\") pod \"neutron-f054-account-create-update-f98s6\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.016071 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l5zg\" (UniqueName: \"kubernetes.io/projected/97386078-2478-4899-be49-4f8bb3019081-kube-api-access-4l5zg\") pod \"neutron-f054-account-create-update-f98s6\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.128987 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.271040 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qjg97"] Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.303349 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0275-account-create-update-sxr85"] Dec 06 10:55:16 crc kubenswrapper[4678]: W1206 10:55:16.324777 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb2d3760e_f3e5_4ae3_93a7_ff1c2f0bf4db.slice/crio-5530085d0e3aacbf9b5f2046929edc0720728d8b536b9c0c10c36edea973de0f WatchSource:0}: Error finding container 5530085d0e3aacbf9b5f2046929edc0720728d8b536b9c0c10c36edea973de0f: Status 404 returned error can't find the container with id 5530085d0e3aacbf9b5f2046929edc0720728d8b536b9c0c10c36edea973de0f Dec 06 10:55:16 crc kubenswrapper[4678]: W1206 10:55:16.326791 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaddc760_3153_40b0_8e44_01e32a9d909f.slice/crio-472b5ff86dd4bcef4c3a1e050eb73b8b404d8ab7096cc28d501eb605dc0dbc24 WatchSource:0}: Error finding container 472b5ff86dd4bcef4c3a1e050eb73b8b404d8ab7096cc28d501eb605dc0dbc24: Status 404 returned error can't find the container with id 472b5ff86dd4bcef4c3a1e050eb73b8b404d8ab7096cc28d501eb605dc0dbc24 Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.503611 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-m54tp"] Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.512412 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qjg97" event={"ID":"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db","Type":"ContainerStarted","Data":"5530085d0e3aacbf9b5f2046929edc0720728d8b536b9c0c10c36edea973de0f"} Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.514155 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0275-account-create-update-sxr85" event={"ID":"daddc760-3153-40b0-8e44-01e32a9d909f","Type":"ContainerStarted","Data":"472b5ff86dd4bcef4c3a1e050eb73b8b404d8ab7096cc28d501eb605dc0dbc24"} Dec 06 10:55:16 crc kubenswrapper[4678]: W1206 10:55:16.517183 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54cdc7e5_a5f1_4b88_b5a6_36ed4fdd08f1.slice/crio-4ec31647f890c464e5b1daaf041d6001c5bb3eb49084f0082e0add0223507dbc WatchSource:0}: Error finding container 4ec31647f890c464e5b1daaf041d6001c5bb3eb49084f0082e0add0223507dbc: Status 404 returned error can't find the container with id 4ec31647f890c464e5b1daaf041d6001c5bb3eb49084f0082e0add0223507dbc Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.709220 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-055f-account-create-update-wmjpx"] Dec 06 10:55:16 crc kubenswrapper[4678]: W1206 10:55:16.731283 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a074406_0662_4e37_8cb5_4f187d687cb8.slice/crio-63a1ab8bbc58745fae02f997ab959944914a0add99463717d69da78d3d452761 WatchSource:0}: Error finding container 63a1ab8bbc58745fae02f997ab959944914a0add99463717d69da78d3d452761: Status 404 returned error can't find the container with id 63a1ab8bbc58745fae02f997ab959944914a0add99463717d69da78d3d452761 Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.746943 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f054-account-create-update-f98s6"] Dec 06 10:55:16 crc kubenswrapper[4678]: W1206 10:55:16.756458 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97386078_2478_4899_be49_4f8bb3019081.slice/crio-b97bd227c2a68dd4abc7f77f8fd63cafc9087b1039f54d27a230d4fec8f9e517 WatchSource:0}: Error finding container b97bd227c2a68dd4abc7f77f8fd63cafc9087b1039f54d27a230d4fec8f9e517: Status 404 returned error can't find the container with id b97bd227c2a68dd4abc7f77f8fd63cafc9087b1039f54d27a230d4fec8f9e517 Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.837696 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mzqdh"] Dec 06 10:55:16 crc kubenswrapper[4678]: I1206 10:55:16.880696 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x9jhf"] Dec 06 10:55:16 crc kubenswrapper[4678]: W1206 10:55:16.903012 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d28bdf3_353f_444d_b401_ff06978869c9.slice/crio-9502f3c129dff6f7c897e7476c3118e8ec696ac9446a29374407570a91413267 WatchSource:0}: Error finding container 9502f3c129dff6f7c897e7476c3118e8ec696ac9446a29374407570a91413267: Status 404 returned error can't find the container with id 9502f3c129dff6f7c897e7476c3118e8ec696ac9446a29374407570a91413267 Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.527394 4678 generic.go:334] "Generic (PLEG): container finished" podID="97386078-2478-4899-be49-4f8bb3019081" containerID="5afb9f55b56a33c076a94c6f6e31da815df753dcc4d5bf5ef8007c7348638920" exitCode=0 Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.527467 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f054-account-create-update-f98s6" event={"ID":"97386078-2478-4899-be49-4f8bb3019081","Type":"ContainerDied","Data":"5afb9f55b56a33c076a94c6f6e31da815df753dcc4d5bf5ef8007c7348638920"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.527513 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f054-account-create-update-f98s6" event={"ID":"97386078-2478-4899-be49-4f8bb3019081","Type":"ContainerStarted","Data":"b97bd227c2a68dd4abc7f77f8fd63cafc9087b1039f54d27a230d4fec8f9e517"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.529421 4678 generic.go:334] "Generic (PLEG): container finished" podID="54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1" containerID="f28de67ad744a935191f8ce3699dbfab7838ac5d52dbbeb3238b39ce433043c7" exitCode=0 Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.529623 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m54tp" event={"ID":"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1","Type":"ContainerDied","Data":"f28de67ad744a935191f8ce3699dbfab7838ac5d52dbbeb3238b39ce433043c7"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.529738 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m54tp" event={"ID":"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1","Type":"ContainerStarted","Data":"4ec31647f890c464e5b1daaf041d6001c5bb3eb49084f0082e0add0223507dbc"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.531962 4678 generic.go:334] "Generic (PLEG): container finished" podID="8a074406-0662-4e37-8cb5-4f187d687cb8" containerID="0007edd53857b7b76507a17ecfaaf83b8926043cffd1737c98f9d930f8127c02" exitCode=0 Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.532023 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-055f-account-create-update-wmjpx" event={"ID":"8a074406-0662-4e37-8cb5-4f187d687cb8","Type":"ContainerDied","Data":"0007edd53857b7b76507a17ecfaaf83b8926043cffd1737c98f9d930f8127c02"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.532221 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-055f-account-create-update-wmjpx" event={"ID":"8a074406-0662-4e37-8cb5-4f187d687cb8","Type":"ContainerStarted","Data":"63a1ab8bbc58745fae02f997ab959944914a0add99463717d69da78d3d452761"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.533375 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mzqdh" event={"ID":"26bbb0e2-894e-4476-b5c8-4af78dedafd8","Type":"ContainerStarted","Data":"46dca4b1b8ac4bdf62b6b0c5a9f12bfb2aef3fe6d2b0c7254d7f6764e274f3f5"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.534734 4678 generic.go:334] "Generic (PLEG): container finished" podID="b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db" containerID="990ce8ae0bd650b55dd57e3a758706a4534166a91d32aa8bb76b17c751b15b51" exitCode=0 Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.534796 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qjg97" event={"ID":"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db","Type":"ContainerDied","Data":"990ce8ae0bd650b55dd57e3a758706a4534166a91d32aa8bb76b17c751b15b51"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.536403 4678 generic.go:334] "Generic (PLEG): container finished" podID="7d28bdf3-353f-444d-b401-ff06978869c9" containerID="c0da59169d40ad44d3181acffa370adebc4a12b3ffad4637228e8e8fdb7e9401" exitCode=0 Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.536507 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9jhf" event={"ID":"7d28bdf3-353f-444d-b401-ff06978869c9","Type":"ContainerDied","Data":"c0da59169d40ad44d3181acffa370adebc4a12b3ffad4637228e8e8fdb7e9401"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.536571 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9jhf" event={"ID":"7d28bdf3-353f-444d-b401-ff06978869c9","Type":"ContainerStarted","Data":"9502f3c129dff6f7c897e7476c3118e8ec696ac9446a29374407570a91413267"} Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.546518 4678 generic.go:334] "Generic (PLEG): container finished" podID="daddc760-3153-40b0-8e44-01e32a9d909f" containerID="b069c241c02034d05ec4f49072ca4fdc939d62f2e0d0127aa4598e6a40a950d1" exitCode=0 Dec 06 10:55:17 crc kubenswrapper[4678]: I1206 10:55:17.546565 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0275-account-create-update-sxr85" event={"ID":"daddc760-3153-40b0-8e44-01e32a9d909f","Type":"ContainerDied","Data":"b069c241c02034d05ec4f49072ca4fdc939d62f2e0d0127aa4598e6a40a950d1"} Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.013424 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.159475 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a074406-0662-4e37-8cb5-4f187d687cb8-operator-scripts\") pod \"8a074406-0662-4e37-8cb5-4f187d687cb8\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.159896 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h488v\" (UniqueName: \"kubernetes.io/projected/8a074406-0662-4e37-8cb5-4f187d687cb8-kube-api-access-h488v\") pod \"8a074406-0662-4e37-8cb5-4f187d687cb8\" (UID: \"8a074406-0662-4e37-8cb5-4f187d687cb8\") " Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.160877 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a074406-0662-4e37-8cb5-4f187d687cb8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a074406-0662-4e37-8cb5-4f187d687cb8" (UID: "8a074406-0662-4e37-8cb5-4f187d687cb8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.167090 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a074406-0662-4e37-8cb5-4f187d687cb8-kube-api-access-h488v" (OuterVolumeSpecName: "kube-api-access-h488v") pod "8a074406-0662-4e37-8cb5-4f187d687cb8" (UID: "8a074406-0662-4e37-8cb5-4f187d687cb8"). InnerVolumeSpecName "kube-api-access-h488v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.261582 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a074406-0662-4e37-8cb5-4f187d687cb8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.261617 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h488v\" (UniqueName: \"kubernetes.io/projected/8a074406-0662-4e37-8cb5-4f187d687cb8-kube-api-access-h488v\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.374555 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.440406 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jlb2j"] Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.440696 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerName="dnsmasq-dns" containerID="cri-o://ae64fa013429922acaccb6ecd8a497f70287113248adc2fc9df366cb096103c2" gracePeriod=10 Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.579060 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-055f-account-create-update-wmjpx" event={"ID":"8a074406-0662-4e37-8cb5-4f187d687cb8","Type":"ContainerDied","Data":"63a1ab8bbc58745fae02f997ab959944914a0add99463717d69da78d3d452761"} Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.579329 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63a1ab8bbc58745fae02f997ab959944914a0add99463717d69da78d3d452761" Dec 06 10:55:19 crc kubenswrapper[4678]: I1206 10:55:19.579260 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-055f-account-create-update-wmjpx" Dec 06 10:55:20 crc kubenswrapper[4678]: I1206 10:55:20.214132 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Dec 06 10:55:20 crc kubenswrapper[4678]: I1206 10:55:20.590623 4678 generic.go:334] "Generic (PLEG): container finished" podID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerID="ae64fa013429922acaccb6ecd8a497f70287113248adc2fc9df366cb096103c2" exitCode=0 Dec 06 10:55:20 crc kubenswrapper[4678]: I1206 10:55:20.590670 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" event={"ID":"4c396d76-0ede-4141-b9be-ffba8a8f8f2d","Type":"ContainerDied","Data":"ae64fa013429922acaccb6ecd8a497f70287113248adc2fc9df366cb096103c2"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.161246 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.169352 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.183799 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.234714 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.276065 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316253 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-operator-scripts\") pod \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316312 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97386078-2478-4899-be49-4f8bb3019081-operator-scripts\") pod \"97386078-2478-4899-be49-4f8bb3019081\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316433 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pglbz\" (UniqueName: \"kubernetes.io/projected/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-kube-api-access-pglbz\") pod \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316526 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsqsk\" (UniqueName: \"kubernetes.io/projected/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-kube-api-access-fsqsk\") pod \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\" (UID: \"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316598 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkvdr\" (UniqueName: \"kubernetes.io/projected/daddc760-3153-40b0-8e44-01e32a9d909f-kube-api-access-xkvdr\") pod \"daddc760-3153-40b0-8e44-01e32a9d909f\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316664 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l5zg\" (UniqueName: \"kubernetes.io/projected/97386078-2478-4899-be49-4f8bb3019081-kube-api-access-4l5zg\") pod \"97386078-2478-4899-be49-4f8bb3019081\" (UID: \"97386078-2478-4899-be49-4f8bb3019081\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316698 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-operator-scripts\") pod \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\" (UID: \"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.316722 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/daddc760-3153-40b0-8e44-01e32a9d909f-operator-scripts\") pod \"daddc760-3153-40b0-8e44-01e32a9d909f\" (UID: \"daddc760-3153-40b0-8e44-01e32a9d909f\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.317526 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db" (UID: "b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.318643 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/daddc760-3153-40b0-8e44-01e32a9d909f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "daddc760-3153-40b0-8e44-01e32a9d909f" (UID: "daddc760-3153-40b0-8e44-01e32a9d909f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.318955 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97386078-2478-4899-be49-4f8bb3019081-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "97386078-2478-4899-be49-4f8bb3019081" (UID: "97386078-2478-4899-be49-4f8bb3019081"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.319383 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1" (UID: "54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.323865 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-kube-api-access-pglbz" (OuterVolumeSpecName: "kube-api-access-pglbz") pod "54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1" (UID: "54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1"). InnerVolumeSpecName "kube-api-access-pglbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.339073 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-kube-api-access-fsqsk" (OuterVolumeSpecName: "kube-api-access-fsqsk") pod "b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db" (UID: "b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db"). InnerVolumeSpecName "kube-api-access-fsqsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.341020 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daddc760-3153-40b0-8e44-01e32a9d909f-kube-api-access-xkvdr" (OuterVolumeSpecName: "kube-api-access-xkvdr") pod "daddc760-3153-40b0-8e44-01e32a9d909f" (UID: "daddc760-3153-40b0-8e44-01e32a9d909f"). InnerVolumeSpecName "kube-api-access-xkvdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.341731 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97386078-2478-4899-be49-4f8bb3019081-kube-api-access-4l5zg" (OuterVolumeSpecName: "kube-api-access-4l5zg") pod "97386078-2478-4899-be49-4f8bb3019081" (UID: "97386078-2478-4899-be49-4f8bb3019081"). InnerVolumeSpecName "kube-api-access-4l5zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.380786 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.418525 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkrg2\" (UniqueName: \"kubernetes.io/projected/7d28bdf3-353f-444d-b401-ff06978869c9-kube-api-access-kkrg2\") pod \"7d28bdf3-353f-444d-b401-ff06978869c9\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.418682 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d28bdf3-353f-444d-b401-ff06978869c9-operator-scripts\") pod \"7d28bdf3-353f-444d-b401-ff06978869c9\" (UID: \"7d28bdf3-353f-444d-b401-ff06978869c9\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419149 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkvdr\" (UniqueName: \"kubernetes.io/projected/daddc760-3153-40b0-8e44-01e32a9d909f-kube-api-access-xkvdr\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419183 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l5zg\" (UniqueName: \"kubernetes.io/projected/97386078-2478-4899-be49-4f8bb3019081-kube-api-access-4l5zg\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419196 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419205 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/daddc760-3153-40b0-8e44-01e32a9d909f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419216 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419224 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/97386078-2478-4899-be49-4f8bb3019081-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419234 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pglbz\" (UniqueName: \"kubernetes.io/projected/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1-kube-api-access-pglbz\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419267 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsqsk\" (UniqueName: \"kubernetes.io/projected/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db-kube-api-access-fsqsk\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.419575 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d28bdf3-353f-444d-b401-ff06978869c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7d28bdf3-353f-444d-b401-ff06978869c9" (UID: "7d28bdf3-353f-444d-b401-ff06978869c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.423703 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d28bdf3-353f-444d-b401-ff06978869c9-kube-api-access-kkrg2" (OuterVolumeSpecName: "kube-api-access-kkrg2") pod "7d28bdf3-353f-444d-b401-ff06978869c9" (UID: "7d28bdf3-353f-444d-b401-ff06978869c9"). InnerVolumeSpecName "kube-api-access-kkrg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.532661 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-sb\") pod \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.532739 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-dns-svc\") pod \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.532800 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-config\") pod \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.532911 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-nb\") pod \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.532939 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ctvr\" (UniqueName: \"kubernetes.io/projected/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-kube-api-access-8ctvr\") pod \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\" (UID: \"4c396d76-0ede-4141-b9be-ffba8a8f8f2d\") " Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.533648 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7d28bdf3-353f-444d-b401-ff06978869c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.534546 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkrg2\" (UniqueName: \"kubernetes.io/projected/7d28bdf3-353f-444d-b401-ff06978869c9-kube-api-access-kkrg2\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.539816 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-kube-api-access-8ctvr" (OuterVolumeSpecName: "kube-api-access-8ctvr") pod "4c396d76-0ede-4141-b9be-ffba8a8f8f2d" (UID: "4c396d76-0ede-4141-b9be-ffba8a8f8f2d"). InnerVolumeSpecName "kube-api-access-8ctvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.571226 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-config" (OuterVolumeSpecName: "config") pod "4c396d76-0ede-4141-b9be-ffba8a8f8f2d" (UID: "4c396d76-0ede-4141-b9be-ffba8a8f8f2d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.576152 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4c396d76-0ede-4141-b9be-ffba8a8f8f2d" (UID: "4c396d76-0ede-4141-b9be-ffba8a8f8f2d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.577920 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4c396d76-0ede-4141-b9be-ffba8a8f8f2d" (UID: "4c396d76-0ede-4141-b9be-ffba8a8f8f2d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.584105 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4c396d76-0ede-4141-b9be-ffba8a8f8f2d" (UID: "4c396d76-0ede-4141-b9be-ffba8a8f8f2d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.611049 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mzqdh" event={"ID":"26bbb0e2-894e-4476-b5c8-4af78dedafd8","Type":"ContainerStarted","Data":"e46c9a15f14bc7ee694c2b6999da21360588452aa5aaf7fb51a13df03e13dce8"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.613812 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qjg97" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.613766 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qjg97" event={"ID":"b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db","Type":"ContainerDied","Data":"5530085d0e3aacbf9b5f2046929edc0720728d8b536b9c0c10c36edea973de0f"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.614010 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5530085d0e3aacbf9b5f2046929edc0720728d8b536b9c0c10c36edea973de0f" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.615461 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x9jhf" event={"ID":"7d28bdf3-353f-444d-b401-ff06978869c9","Type":"ContainerDied","Data":"9502f3c129dff6f7c897e7476c3118e8ec696ac9446a29374407570a91413267"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.615617 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9502f3c129dff6f7c897e7476c3118e8ec696ac9446a29374407570a91413267" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.616039 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x9jhf" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.616789 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0275-account-create-update-sxr85" event={"ID":"daddc760-3153-40b0-8e44-01e32a9d909f","Type":"ContainerDied","Data":"472b5ff86dd4bcef4c3a1e050eb73b8b404d8ab7096cc28d501eb605dc0dbc24"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.616914 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="472b5ff86dd4bcef4c3a1e050eb73b8b404d8ab7096cc28d501eb605dc0dbc24" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.616844 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0275-account-create-update-sxr85" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.619108 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" event={"ID":"4c396d76-0ede-4141-b9be-ffba8a8f8f2d","Type":"ContainerDied","Data":"3de61fa0da586a30c3a305cb0d5c57592fdd92000e46c664de96d777d112ce4f"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.619151 4678 scope.go:117] "RemoveContainer" containerID="ae64fa013429922acaccb6ecd8a497f70287113248adc2fc9df366cb096103c2" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.619117 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-jlb2j" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.621094 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f054-account-create-update-f98s6" event={"ID":"97386078-2478-4899-be49-4f8bb3019081","Type":"ContainerDied","Data":"b97bd227c2a68dd4abc7f77f8fd63cafc9087b1039f54d27a230d4fec8f9e517"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.621119 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b97bd227c2a68dd4abc7f77f8fd63cafc9087b1039f54d27a230d4fec8f9e517" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.621177 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f054-account-create-update-f98s6" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.641078 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-m54tp" event={"ID":"54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1","Type":"ContainerDied","Data":"4ec31647f890c464e5b1daaf041d6001c5bb3eb49084f0082e0add0223507dbc"} Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.641115 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ec31647f890c464e5b1daaf041d6001c5bb3eb49084f0082e0add0223507dbc" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.641220 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-m54tp" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.646198 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.646759 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.646786 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ctvr\" (UniqueName: \"kubernetes.io/projected/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-kube-api-access-8ctvr\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.646802 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.646816 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4c396d76-0ede-4141-b9be-ffba8a8f8f2d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.647633 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-mzqdh" podStartSLOduration=2.4453428649999998 podStartE2EDuration="7.647615679s" podCreationTimestamp="2025-12-06 10:55:15 +0000 UTC" firstStartedPulling="2025-12-06 10:55:16.853367654 +0000 UTC m=+1121.696799093" lastFinishedPulling="2025-12-06 10:55:22.055640468 +0000 UTC m=+1126.899071907" observedRunningTime="2025-12-06 10:55:22.629314241 +0000 UTC m=+1127.472745680" watchObservedRunningTime="2025-12-06 10:55:22.647615679 +0000 UTC m=+1127.491047118" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.668253 4678 scope.go:117] "RemoveContainer" containerID="e8aa9719bf55fc0c42d1c39280635cfceb8166de99c0a5f8a8c0ee0ae11053f7" Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.707610 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jlb2j"] Dec 06 10:55:22 crc kubenswrapper[4678]: I1206 10:55:22.715260 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-jlb2j"] Dec 06 10:55:23 crc kubenswrapper[4678]: I1206 10:55:23.488228 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" path="/var/lib/kubelet/pods/4c396d76-0ede-4141-b9be-ffba8a8f8f2d/volumes" Dec 06 10:55:26 crc kubenswrapper[4678]: I1206 10:55:26.677639 4678 generic.go:334] "Generic (PLEG): container finished" podID="26bbb0e2-894e-4476-b5c8-4af78dedafd8" containerID="e46c9a15f14bc7ee694c2b6999da21360588452aa5aaf7fb51a13df03e13dce8" exitCode=0 Dec 06 10:55:26 crc kubenswrapper[4678]: I1206 10:55:26.677965 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mzqdh" event={"ID":"26bbb0e2-894e-4476-b5c8-4af78dedafd8","Type":"ContainerDied","Data":"e46c9a15f14bc7ee694c2b6999da21360588452aa5aaf7fb51a13df03e13dce8"} Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.029410 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.144327 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-combined-ca-bundle\") pod \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.144772 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsqp4\" (UniqueName: \"kubernetes.io/projected/26bbb0e2-894e-4476-b5c8-4af78dedafd8-kube-api-access-dsqp4\") pod \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.144882 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-config-data\") pod \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\" (UID: \"26bbb0e2-894e-4476-b5c8-4af78dedafd8\") " Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.151771 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26bbb0e2-894e-4476-b5c8-4af78dedafd8-kube-api-access-dsqp4" (OuterVolumeSpecName: "kube-api-access-dsqp4") pod "26bbb0e2-894e-4476-b5c8-4af78dedafd8" (UID: "26bbb0e2-894e-4476-b5c8-4af78dedafd8"). InnerVolumeSpecName "kube-api-access-dsqp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.188545 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26bbb0e2-894e-4476-b5c8-4af78dedafd8" (UID: "26bbb0e2-894e-4476-b5c8-4af78dedafd8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.191727 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-config-data" (OuterVolumeSpecName: "config-data") pod "26bbb0e2-894e-4476-b5c8-4af78dedafd8" (UID: "26bbb0e2-894e-4476-b5c8-4af78dedafd8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.246791 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.246841 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsqp4\" (UniqueName: \"kubernetes.io/projected/26bbb0e2-894e-4476-b5c8-4af78dedafd8-kube-api-access-dsqp4\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.246862 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26bbb0e2-894e-4476-b5c8-4af78dedafd8-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.704857 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mzqdh" event={"ID":"26bbb0e2-894e-4476-b5c8-4af78dedafd8","Type":"ContainerDied","Data":"46dca4b1b8ac4bdf62b6b0c5a9f12bfb2aef3fe6d2b0c7254d7f6764e274f3f5"} Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.704955 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46dca4b1b8ac4bdf62b6b0c5a9f12bfb2aef3fe6d2b0c7254d7f6764e274f3f5" Dec 06 10:55:28 crc kubenswrapper[4678]: I1206 10:55:28.705073 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mzqdh" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.006630 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-g6ffn"] Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.006966 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97386078-2478-4899-be49-4f8bb3019081" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.006982 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="97386078-2478-4899-be49-4f8bb3019081" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.006996 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a074406-0662-4e37-8cb5-4f187d687cb8" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007003 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a074406-0662-4e37-8cb5-4f187d687cb8" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.007014 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007020 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.007029 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007035 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.007041 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daddc760-3153-40b0-8e44-01e32a9d909f" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007046 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="daddc760-3153-40b0-8e44-01e32a9d909f" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.007056 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26bbb0e2-894e-4476-b5c8-4af78dedafd8" containerName="keystone-db-sync" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007062 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="26bbb0e2-894e-4476-b5c8-4af78dedafd8" containerName="keystone-db-sync" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.007079 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerName="init" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007085 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerName="init" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.007102 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d28bdf3-353f-444d-b401-ff06978869c9" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007107 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d28bdf3-353f-444d-b401-ff06978869c9" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.007117 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerName="dnsmasq-dns" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007122 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerName="dnsmasq-dns" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007263 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c396d76-0ede-4141-b9be-ffba8a8f8f2d" containerName="dnsmasq-dns" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007277 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007288 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="26bbb0e2-894e-4476-b5c8-4af78dedafd8" containerName="keystone-db-sync" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007298 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="97386078-2478-4899-be49-4f8bb3019081" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007306 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d28bdf3-353f-444d-b401-ff06978869c9" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007330 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="daddc760-3153-40b0-8e44-01e32a9d909f" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007344 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1" containerName="mariadb-database-create" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.007355 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a074406-0662-4e37-8cb5-4f187d687cb8" containerName="mariadb-account-create-update" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.008190 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.040401 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-g6ffn"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.067439 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrsz9\" (UniqueName: \"kubernetes.io/projected/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-kube-api-access-rrsz9\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.071854 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.071968 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.072050 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.072148 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.072926 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.085156 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-62hxf"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.105756 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.133646 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.133922 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z5rpk" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.134089 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.133779 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.134350 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.142207 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-62hxf"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175165 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175228 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25qcr\" (UniqueName: \"kubernetes.io/projected/135a09b4-e309-47bc-8905-180c0060a639-kube-api-access-25qcr\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175283 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrsz9\" (UniqueName: \"kubernetes.io/projected/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-kube-api-access-rrsz9\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175308 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-fernet-keys\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175326 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-scripts\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175361 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175381 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-combined-ca-bundle\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175406 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175422 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-credential-keys\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175442 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-config-data\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175463 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.175504 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.176321 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.176910 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.180647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.181212 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.185289 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.218349 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrsz9\" (UniqueName: \"kubernetes.io/projected/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-kube-api-access-rrsz9\") pod \"dnsmasq-dns-bbf5cc879-g6ffn\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.276948 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25qcr\" (UniqueName: \"kubernetes.io/projected/135a09b4-e309-47bc-8905-180c0060a639-kube-api-access-25qcr\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.277033 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-fernet-keys\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.277060 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-scripts\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.277103 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-combined-ca-bundle\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.277135 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-credential-keys\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.277159 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-config-data\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.283330 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-config-data\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.289865 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-scripts\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.290284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-fernet-keys\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.298167 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-combined-ca-bundle\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.306947 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-credential-keys\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.328080 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.385017 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25qcr\" (UniqueName: \"kubernetes.io/projected/135a09b4-e309-47bc-8905-180c0060a639-kube-api-access-25qcr\") pod \"keystone-bootstrap-62hxf\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.474269 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.509367 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.509430 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.585785 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c6f6bc877-zxzjz"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.599710 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.640305 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-c448f"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.648046 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.650510 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.651691 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.651858 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.651949 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-7dxlf" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.695673 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-scripts\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.695740 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66418cf7-c293-4ddc-948c-60666d9fd1c4-logs\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.695769 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-config-data\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.695793 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwchh\" (UniqueName: \"kubernetes.io/projected/66418cf7-c293-4ddc-948c-60666d9fd1c4-kube-api-access-rwchh\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.695814 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66418cf7-c293-4ddc-948c-60666d9fd1c4-horizon-secret-key\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: W1206 10:55:29.704068 4678 reflector.go:561] object-"openstack"/"cinder-cinder-dockercfg-g5rcq": failed to list *v1.Secret: secrets "cinder-cinder-dockercfg-g5rcq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.704119 4678 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cinder-cinder-dockercfg-g5rcq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cinder-cinder-dockercfg-g5rcq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 10:55:29 crc kubenswrapper[4678]: W1206 10:55:29.704163 4678 reflector.go:561] object-"openstack"/"cinder-scripts": failed to list *v1.Secret: secrets "cinder-scripts" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.704176 4678 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cinder-scripts\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cinder-scripts\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 10:55:29 crc kubenswrapper[4678]: W1206 10:55:29.704213 4678 reflector.go:561] object-"openstack"/"cinder-config-data": failed to list *v1.Secret: secrets "cinder-config-data" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 06 10:55:29 crc kubenswrapper[4678]: E1206 10:55:29.704223 4678 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cinder-config-data\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cinder-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.731826 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c6f6bc877-zxzjz"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.753935 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-c448f"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799442 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-scripts\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799506 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c315badb-6f3b-408f-b68f-b359cea62f98-etc-machine-id\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799544 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-db-sync-config-data\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799575 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-config-data\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799598 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9qtz\" (UniqueName: \"kubernetes.io/projected/c315badb-6f3b-408f-b68f-b359cea62f98-kube-api-access-q9qtz\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799615 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66418cf7-c293-4ddc-948c-60666d9fd1c4-logs\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799640 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-combined-ca-bundle\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-config-data\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799685 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwchh\" (UniqueName: \"kubernetes.io/projected/66418cf7-c293-4ddc-948c-60666d9fd1c4-kube-api-access-rwchh\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799701 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-scripts\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.799720 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66418cf7-c293-4ddc-948c-60666d9fd1c4-horizon-secret-key\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.800875 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66418cf7-c293-4ddc-948c-60666d9fd1c4-logs\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.801852 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-scripts\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.802103 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-config-data\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.812831 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66418cf7-c293-4ddc-948c-60666d9fd1c4-horizon-secret-key\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.900600 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c315badb-6f3b-408f-b68f-b359cea62f98-etc-machine-id\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.900653 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-db-sync-config-data\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.900689 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-config-data\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.900707 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9qtz\" (UniqueName: \"kubernetes.io/projected/c315badb-6f3b-408f-b68f-b359cea62f98-kube-api-access-q9qtz\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.900735 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-combined-ca-bundle\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.900771 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-scripts\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.900944 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c315badb-6f3b-408f-b68f-b359cea62f98-etc-machine-id\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.907966 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-combined-ca-bundle\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.918872 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-cvhc9"] Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.920134 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.925085 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwchh\" (UniqueName: \"kubernetes.io/projected/66418cf7-c293-4ddc-948c-60666d9fd1c4-kube-api-access-rwchh\") pod \"horizon-c6f6bc877-zxzjz\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:29 crc kubenswrapper[4678]: I1206 10:55:29.963200 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.003541 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-config-data\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.003617 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-logs\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.003707 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-combined-ca-bundle\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.003742 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-scripts\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.003777 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k456h\" (UniqueName: \"kubernetes.io/projected/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-kube-api-access-k456h\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.014732 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.022766 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.023004 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pr6nm" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.078333 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cvhc9"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.083234 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9qtz\" (UniqueName: \"kubernetes.io/projected/c315badb-6f3b-408f-b68f-b359cea62f98-kube-api-access-q9qtz\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.105387 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-logs\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.105519 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-combined-ca-bundle\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.105553 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-scripts\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.105579 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k456h\" (UniqueName: \"kubernetes.io/projected/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-kube-api-access-k456h\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.105642 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-config-data\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.105942 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-logs\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.118284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-scripts\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.119111 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-config-data\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.128280 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-combined-ca-bundle\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.132548 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-lmpj6"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.133741 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.148964 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.149141 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.149234 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8hhm2" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.178650 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k456h\" (UniqueName: \"kubernetes.io/projected/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-kube-api-access-k456h\") pod \"placement-db-sync-cvhc9\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.206828 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-config\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.207184 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-combined-ca-bundle\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.207304 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjs4j\" (UniqueName: \"kubernetes.io/projected/cca26a10-9756-4a74-8a33-688f36c21b46-kube-api-access-xjs4j\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.265793 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-g6ffn"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.272480 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xj2n5"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.273675 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.279923 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7kv78" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.296659 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.297454 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lmpj6"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.300818 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cvhc9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.321088 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjs4j\" (UniqueName: \"kubernetes.io/projected/cca26a10-9756-4a74-8a33-688f36c21b46-kube-api-access-xjs4j\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.321179 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-config\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.321227 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-combined-ca-bundle\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.350229 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-config\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.369401 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-combined-ca-bundle\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.394677 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xj2n5"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.398828 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjs4j\" (UniqueName: \"kubernetes.io/projected/cca26a10-9756-4a74-8a33-688f36c21b46-kube-api-access-xjs4j\") pod \"neutron-db-sync-lmpj6\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.435747 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-db-sync-config-data\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.435981 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-combined-ca-bundle\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.436010 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2hrb\" (UniqueName: \"kubernetes.io/projected/3a95e322-b0b3-4083-81a3-187b579c53f4-kube-api-access-n2hrb\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.447568 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b8b488c65-nmwlp"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.448963 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.495929 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-tfqd9"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.500068 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.500678 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541353 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-logs\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541408 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-db-sync-config-data\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541431 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-combined-ca-bundle\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541448 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-scripts\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541467 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2hrb\" (UniqueName: \"kubernetes.io/projected/3a95e322-b0b3-4083-81a3-187b579c53f4-kube-api-access-n2hrb\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541527 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gmdh\" (UniqueName: \"kubernetes.io/projected/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-kube-api-access-2gmdh\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541557 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-horizon-secret-key\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.541592 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-config-data\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.547052 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-g6ffn"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.551338 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-db-sync-config-data\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.555503 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-g5rcq" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.558524 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.559877 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.564872 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.565205 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xrkxl" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.565232 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-combined-ca-bundle\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.565408 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.565551 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.566403 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-tfqd9"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.565767 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2hrb\" (UniqueName: \"kubernetes.io/projected/3a95e322-b0b3-4083-81a3-187b579c53f4-kube-api-access-n2hrb\") pod \"barbican-db-sync-xj2n5\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.587572 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b8b488c65-nmwlp"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.602527 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.607394 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-db-sync-config-data\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.620178 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.627831 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-config-data\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.649242 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.664500 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-horizon-secret-key\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.664608 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.664786 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-config-data\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.664846 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.664878 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.664908 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.665146 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.665246 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj6pm\" (UniqueName: \"kubernetes.io/projected/ced0ddf7-759e-4cb9-9516-02174759fddf-kube-api-access-kj6pm\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.665474 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-config\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.665560 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2qvx\" (UniqueName: \"kubernetes.io/projected/6316f793-eac2-4be6-8cc1-63701fc9ca88-kube-api-access-j2qvx\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.665739 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-logs\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.665766 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-logs\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.665793 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.666049 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.666080 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-scripts\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.668909 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.668943 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.669023 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gmdh\" (UniqueName: \"kubernetes.io/projected/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-kube-api-access-2gmdh\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.669212 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.673437 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.676481 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-horizon-secret-key\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.682320 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-scripts\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.683335 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-logs\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.697756 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-config-data\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.714016 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gmdh\" (UniqueName: \"kubernetes.io/projected/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-kube-api-access-2gmdh\") pod \"horizon-7b8b488c65-nmwlp\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.723335 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.740179 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.747333 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 10:55:30 crc kubenswrapper[4678]: I1206 10:55:30.751539 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-scripts\") pod \"cinder-db-sync-c448f\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825498 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-logs\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825554 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-scripts\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825584 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825611 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825687 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825733 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825758 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825853 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825899 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-config-data\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.825953 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826003 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826076 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826095 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826112 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826207 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj6pm\" (UniqueName: \"kubernetes.io/projected/ced0ddf7-759e-4cb9-9516-02174759fddf-kube-api-access-kj6pm\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826238 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826274 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-config\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826297 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826332 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlmpt\" (UniqueName: \"kubernetes.io/projected/e92229b3-87c1-45dd-986b-a362109b6a83-kube-api-access-vlmpt\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826367 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-logs\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.826395 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2qvx\" (UniqueName: \"kubernetes.io/projected/6316f793-eac2-4be6-8cc1-63701fc9ca88-kube-api-access-j2qvx\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.827652 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-logs\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.832440 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.833889 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.834373 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.836802 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.837312 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.839766 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.840467 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.842967 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-config\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.843272 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.863432 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.866682 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.892199 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.905454 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.906117 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj6pm\" (UniqueName: \"kubernetes.io/projected/ced0ddf7-759e-4cb9-9516-02174759fddf-kube-api-access-kj6pm\") pod \"dnsmasq-dns-56df8fb6b7-tfqd9\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.908208 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.909047 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2qvx\" (UniqueName: \"kubernetes.io/projected/6316f793-eac2-4be6-8cc1-63701fc9ca88-kube-api-access-j2qvx\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.912373 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c448f" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933413 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933474 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-config-data\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933518 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933581 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933605 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933627 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlmpt\" (UniqueName: \"kubernetes.io/projected/e92229b3-87c1-45dd-986b-a362109b6a83-kube-api-access-vlmpt\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933649 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-logs\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.933679 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-scripts\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.947009 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.947109 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.956798 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.957887 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-logs\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.961036 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.975665 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-62hxf"] Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.975704 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-62hxf" event={"ID":"135a09b4-e309-47bc-8905-180c0060a639","Type":"ContainerStarted","Data":"73ad18e37c074b820a85eb6aad94deb6d01cd99775bafa9ceae61e03e7a69ed5"} Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.975798 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.996421 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-scripts\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.996770 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.996994 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.997111 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:30.998960 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" event={"ID":"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14","Type":"ContainerStarted","Data":"51c247730006f4753d23500dff99f1d7208e7d46f2cf2dcd33981e57bdf5d1cd"} Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.005148 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.007227 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.008047 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-config-data\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.024660 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlmpt\" (UniqueName: \"kubernetes.io/projected/e92229b3-87c1-45dd-986b-a362109b6a83-kube-api-access-vlmpt\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.062705 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.076992 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.136361 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c6f6bc877-zxzjz"] Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.138581 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.138615 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-log-httpd\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.138643 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvf7v\" (UniqueName: \"kubernetes.io/projected/3854c241-241b-49c9-a095-bac632e89d5b-kube-api-access-gvf7v\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.138664 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-run-httpd\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.138737 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.138762 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-config-data\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.138799 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-scripts\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: W1206 10:55:31.145816 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66418cf7_c293_4ddc_948c_60666d9fd1c4.slice/crio-fd8f5f6c7062b21b224751e56d6b035c9f50ec25d55c1bdb6d5a7fb2a1472599 WatchSource:0}: Error finding container fd8f5f6c7062b21b224751e56d6b035c9f50ec25d55c1bdb6d5a7fb2a1472599: Status 404 returned error can't find the container with id fd8f5f6c7062b21b224751e56d6b035c9f50ec25d55c1bdb6d5a7fb2a1472599 Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.248689 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.249021 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-config-data\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.249104 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-scripts\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.249160 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.249174 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-log-httpd\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.249213 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvf7v\" (UniqueName: \"kubernetes.io/projected/3854c241-241b-49c9-a095-bac632e89d5b-kube-api-access-gvf7v\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.249243 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-run-httpd\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.252519 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-run-httpd\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.265329 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-log-httpd\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.267395 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.267881 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-scripts\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.272552 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.284539 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-config-data\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.284860 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.291290 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvf7v\" (UniqueName: \"kubernetes.io/projected/3854c241-241b-49c9-a095-bac632e89d5b-kube-api-access-gvf7v\") pod \"ceilometer-0\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.353028 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:31.374438 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.012341 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6f6bc877-zxzjz" event={"ID":"66418cf7-c293-4ddc-948c-60666d9fd1c4","Type":"ContainerStarted","Data":"fd8f5f6c7062b21b224751e56d6b035c9f50ec25d55c1bdb6d5a7fb2a1472599"} Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.014204 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-62hxf" event={"ID":"135a09b4-e309-47bc-8905-180c0060a639","Type":"ContainerStarted","Data":"318c3bef97a83b0cbc09c87af9eff979f07e306aa5f9bd07e6ca3aad5c757815"} Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.017274 4678 generic.go:334] "Generic (PLEG): container finished" podID="78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" containerID="cac1d316c1f160a3eca7cbd6f7b89a11ae08dc18b4d7890e8050e2ec4601bc65" exitCode=0 Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.017316 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" event={"ID":"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14","Type":"ContainerDied","Data":"cac1d316c1f160a3eca7cbd6f7b89a11ae08dc18b4d7890e8050e2ec4601bc65"} Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.047115 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-62hxf" podStartSLOduration=3.047097671 podStartE2EDuration="3.047097671s" podCreationTimestamp="2025-12-06 10:55:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:55:32.042123084 +0000 UTC m=+1136.885554523" watchObservedRunningTime="2025-12-06 10:55:32.047097671 +0000 UTC m=+1136.890529110" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.413561 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cvhc9"] Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.670955 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.802320 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-sb\") pod \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.802367 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-nb\") pod \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.802438 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-swift-storage-0\") pod \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.802533 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrsz9\" (UniqueName: \"kubernetes.io/projected/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-kube-api-access-rrsz9\") pod \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.802694 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config\") pod \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.802751 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-svc\") pod \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.841035 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-kube-api-access-rrsz9" (OuterVolumeSpecName: "kube-api-access-rrsz9") pod "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" (UID: "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14"). InnerVolumeSpecName "kube-api-access-rrsz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.881274 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" (UID: "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.882439 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" (UID: "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.884932 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" (UID: "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.906221 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.906255 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrsz9\" (UniqueName: \"kubernetes.io/projected/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-kube-api-access-rrsz9\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.906269 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.906278 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:32 crc kubenswrapper[4678]: E1206 10:55:32.917951 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config podName:78ebf1bd-0a1f-45ff-96bb-788ca4f50d14 nodeName:}" failed. No retries permitted until 2025-12-06 10:55:33.417924707 +0000 UTC m=+1138.261356146 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config") pod "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" (UID: "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14") : error deleting /var/lib/kubelet/pods/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14/volume-subpaths: remove /var/lib/kubelet/pods/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14/volume-subpaths: no such file or directory Dec 06 10:55:32 crc kubenswrapper[4678]: I1206 10:55:32.918408 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" (UID: "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.025108 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.051885 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" event={"ID":"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14","Type":"ContainerDied","Data":"51c247730006f4753d23500dff99f1d7208e7d46f2cf2dcd33981e57bdf5d1cd"} Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.051943 4678 scope.go:117] "RemoveContainer" containerID="cac1d316c1f160a3eca7cbd6f7b89a11ae08dc18b4d7890e8050e2ec4601bc65" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.052083 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-g6ffn" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.060872 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cvhc9" event={"ID":"8fbf7713-1e99-4a84-87d0-82803fb8d0e4","Type":"ContainerStarted","Data":"84108bcfa8b30b7cf8f509396c6e6218e64d9877765413ceffc107810fc3c875"} Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.128062 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b8b488c65-nmwlp"] Dec 06 10:55:33 crc kubenswrapper[4678]: W1206 10:55:33.146823 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13c9e9db_8842_4412_9d09_4f8fdb8a39cf.slice/crio-8f2c5a23b1403337b829e5345838fd2dfe9bfa318b04e4a83f38ec53a40278e5 WatchSource:0}: Error finding container 8f2c5a23b1403337b829e5345838fd2dfe9bfa318b04e4a83f38ec53a40278e5: Status 404 returned error can't find the container with id 8f2c5a23b1403337b829e5345838fd2dfe9bfa318b04e4a83f38ec53a40278e5 Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.165117 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.189671 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-c448f"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.259209 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xj2n5"] Dec 06 10:55:33 crc kubenswrapper[4678]: W1206 10:55:33.266581 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a95e322_b0b3_4083_81a3_187b579c53f4.slice/crio-aa5671680bcb47d6bb36e6c6eb9c4f9f4dd5940602c8bfeb4154accfd85c0181 WatchSource:0}: Error finding container aa5671680bcb47d6bb36e6c6eb9c4f9f4dd5940602c8bfeb4154accfd85c0181: Status 404 returned error can't find the container with id aa5671680bcb47d6bb36e6c6eb9c4f9f4dd5940602c8bfeb4154accfd85c0181 Dec 06 10:55:33 crc kubenswrapper[4678]: W1206 10:55:33.278374 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcca26a10_9756_4a74_8a33_688f36c21b46.slice/crio-9104a188fd88f2622173de3cdebd2e46a283f659f5e2e684e7b39449117c926f WatchSource:0}: Error finding container 9104a188fd88f2622173de3cdebd2e46a283f659f5e2e684e7b39449117c926f: Status 404 returned error can't find the container with id 9104a188fd88f2622173de3cdebd2e46a283f659f5e2e684e7b39449117c926f Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.286809 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-tfqd9"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.299669 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lmpj6"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.422567 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.439216 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config\") pod \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\" (UID: \"78ebf1bd-0a1f-45ff-96bb-788ca4f50d14\") " Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.439764 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config" (OuterVolumeSpecName: "config") pod "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" (UID: "78ebf1bd-0a1f-45ff-96bb-788ca4f50d14"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.533422 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c6f6bc877-zxzjz"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.541307 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.569426 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.598517 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-dbfccf5c9-tfx4w"] Dec 06 10:55:33 crc kubenswrapper[4678]: E1206 10:55:33.598872 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" containerName="init" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.598884 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" containerName="init" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.599088 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" containerName="init" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.599918 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.619504 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.630892 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dbfccf5c9-tfx4w"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.642293 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-scripts\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.642342 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6faea971-f662-49d1-b48c-5743f127d7b5-horizon-secret-key\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.642363 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-config-data\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.642392 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6faea971-f662-49d1-b48c-5743f127d7b5-logs\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.642501 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcw8b\" (UniqueName: \"kubernetes.io/projected/6faea971-f662-49d1-b48c-5743f127d7b5-kube-api-access-fcw8b\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.657400 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.744159 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6faea971-f662-49d1-b48c-5743f127d7b5-horizon-secret-key\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.744223 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-config-data\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.744263 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6faea971-f662-49d1-b48c-5743f127d7b5-logs\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.744405 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcw8b\" (UniqueName: \"kubernetes.io/projected/6faea971-f662-49d1-b48c-5743f127d7b5-kube-api-access-fcw8b\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.744462 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-scripts\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.745387 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-scripts\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.745692 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6faea971-f662-49d1-b48c-5743f127d7b5-logs\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.745932 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-config-data\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.755954 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6faea971-f662-49d1-b48c-5743f127d7b5-horizon-secret-key\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.795238 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcw8b\" (UniqueName: \"kubernetes.io/projected/6faea971-f662-49d1-b48c-5743f127d7b5-kube-api-access-fcw8b\") pod \"horizon-dbfccf5c9-tfx4w\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.881268 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-g6ffn"] Dec 06 10:55:33 crc kubenswrapper[4678]: I1206 10:55:33.929936 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-g6ffn"] Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.019613 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.073571 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.080186 4678 generic.go:334] "Generic (PLEG): container finished" podID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerID="64d4ce6c62f1f0b60b8cccb02adf9ff56c28776a70da86031032edf2db0f5ac9" exitCode=0 Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.080460 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" event={"ID":"ced0ddf7-759e-4cb9-9516-02174759fddf","Type":"ContainerDied","Data":"64d4ce6c62f1f0b60b8cccb02adf9ff56c28776a70da86031032edf2db0f5ac9"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.080509 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" event={"ID":"ced0ddf7-759e-4cb9-9516-02174759fddf","Type":"ContainerStarted","Data":"eca2934a21b03f93da8b56dcebf47aeb801a9e13dcb0cbdd3c397ac04aaeb611"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.090123 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj2n5" event={"ID":"3a95e322-b0b3-4083-81a3-187b579c53f4","Type":"ContainerStarted","Data":"aa5671680bcb47d6bb36e6c6eb9c4f9f4dd5940602c8bfeb4154accfd85c0181"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.145986 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c448f" event={"ID":"c315badb-6f3b-408f-b68f-b359cea62f98","Type":"ContainerStarted","Data":"9902cb8efe854befd74db740f5f547002aeffb49ced787373b8dd46975519977"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.154991 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerStarted","Data":"3330d196c0ffa687cce8d8c66c958efeddedbdc4d75d93d46b3393cd0f75c6eb"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.193478 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6316f793-eac2-4be6-8cc1-63701fc9ca88","Type":"ContainerStarted","Data":"d6a8e5560e3e57ca1b984af0c58f0de3508feebfdcf09071b1914eb04f9313e0"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.289165 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b488c65-nmwlp" event={"ID":"13c9e9db-8842-4412-9d09-4f8fdb8a39cf","Type":"ContainerStarted","Data":"8f2c5a23b1403337b829e5345838fd2dfe9bfa318b04e4a83f38ec53a40278e5"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.313137 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lmpj6" event={"ID":"cca26a10-9756-4a74-8a33-688f36c21b46","Type":"ContainerStarted","Data":"3f0590eb23cf2fba33a719d3ba3fb99b45e39f94ba69a13180224a0a50bc72e2"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.313180 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lmpj6" event={"ID":"cca26a10-9756-4a74-8a33-688f36c21b46","Type":"ContainerStarted","Data":"9104a188fd88f2622173de3cdebd2e46a283f659f5e2e684e7b39449117c926f"} Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.807300 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-lmpj6" podStartSLOduration=5.807284299 podStartE2EDuration="5.807284299s" podCreationTimestamp="2025-12-06 10:55:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:55:34.349845637 +0000 UTC m=+1139.193277076" watchObservedRunningTime="2025-12-06 10:55:34.807284299 +0000 UTC m=+1139.650715738" Dec 06 10:55:34 crc kubenswrapper[4678]: I1206 10:55:34.810857 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-dbfccf5c9-tfx4w"] Dec 06 10:55:34 crc kubenswrapper[4678]: W1206 10:55:34.866693 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6faea971_f662_49d1_b48c_5743f127d7b5.slice/crio-ed587d7c0cc1a4e3a0047c2e65b4e67cb65ed7cfc5a2576b4c90599f888ddca7 WatchSource:0}: Error finding container ed587d7c0cc1a4e3a0047c2e65b4e67cb65ed7cfc5a2576b4c90599f888ddca7: Status 404 returned error can't find the container with id ed587d7c0cc1a4e3a0047c2e65b4e67cb65ed7cfc5a2576b4c90599f888ddca7 Dec 06 10:55:35 crc kubenswrapper[4678]: I1206 10:55:35.393067 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbfccf5c9-tfx4w" event={"ID":"6faea971-f662-49d1-b48c-5743f127d7b5","Type":"ContainerStarted","Data":"ed587d7c0cc1a4e3a0047c2e65b4e67cb65ed7cfc5a2576b4c90599f888ddca7"} Dec 06 10:55:35 crc kubenswrapper[4678]: I1206 10:55:35.422787 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6316f793-eac2-4be6-8cc1-63701fc9ca88","Type":"ContainerStarted","Data":"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda"} Dec 06 10:55:35 crc kubenswrapper[4678]: I1206 10:55:35.444721 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" event={"ID":"ced0ddf7-759e-4cb9-9516-02174759fddf","Type":"ContainerStarted","Data":"1b5d79aec9cd09780fce89be3ee6ffb3ee057a47e1e10ac244770d45a54a903d"} Dec 06 10:55:35 crc kubenswrapper[4678]: I1206 10:55:35.446206 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:35 crc kubenswrapper[4678]: I1206 10:55:35.466386 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e92229b3-87c1-45dd-986b-a362109b6a83","Type":"ContainerStarted","Data":"20419c2302600717e96c185bb87fe817e112f76fc9554417f4cf54fd2ff44e1f"} Dec 06 10:55:35 crc kubenswrapper[4678]: I1206 10:55:35.506338 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78ebf1bd-0a1f-45ff-96bb-788ca4f50d14" path="/var/lib/kubelet/pods/78ebf1bd-0a1f-45ff-96bb-788ca4f50d14/volumes" Dec 06 10:55:35 crc kubenswrapper[4678]: I1206 10:55:35.525352 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" podStartSLOduration=5.5253381390000005 podStartE2EDuration="5.525338139s" podCreationTimestamp="2025-12-06 10:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:55:35.508224976 +0000 UTC m=+1140.351656415" watchObservedRunningTime="2025-12-06 10:55:35.525338139 +0000 UTC m=+1140.368769578" Dec 06 10:55:36 crc kubenswrapper[4678]: I1206 10:55:36.506775 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e92229b3-87c1-45dd-986b-a362109b6a83","Type":"ContainerStarted","Data":"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96"} Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.524828 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6316f793-eac2-4be6-8cc1-63701fc9ca88","Type":"ContainerStarted","Data":"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3"} Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.524933 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-log" containerID="cri-o://60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda" gracePeriod=30 Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.525176 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-httpd" containerID="cri-o://35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3" gracePeriod=30 Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.536063 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-log" containerID="cri-o://8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96" gracePeriod=30 Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.536296 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e92229b3-87c1-45dd-986b-a362109b6a83","Type":"ContainerStarted","Data":"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6"} Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.536349 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-httpd" containerID="cri-o://aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6" gracePeriod=30 Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.554249 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.554230886 podStartE2EDuration="7.554230886s" podCreationTimestamp="2025-12-06 10:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:55:37.552057872 +0000 UTC m=+1142.395489311" watchObservedRunningTime="2025-12-06 10:55:37.554230886 +0000 UTC m=+1142.397662315" Dec 06 10:55:37 crc kubenswrapper[4678]: I1206 10:55:37.587272 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.587253009 podStartE2EDuration="7.587253009s" podCreationTimestamp="2025-12-06 10:55:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:55:37.584139327 +0000 UTC m=+1142.427570766" watchObservedRunningTime="2025-12-06 10:55:37.587253009 +0000 UTC m=+1142.430684448" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.472621 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578051 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-config-data\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578138 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-logs\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578154 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-httpd-run\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578214 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578235 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2qvx\" (UniqueName: \"kubernetes.io/projected/6316f793-eac2-4be6-8cc1-63701fc9ca88-kube-api-access-j2qvx\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578257 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-combined-ca-bundle\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578293 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-internal-tls-certs\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578371 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-scripts\") pod \"6316f793-eac2-4be6-8cc1-63701fc9ca88\" (UID: \"6316f793-eac2-4be6-8cc1-63701fc9ca88\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.578930 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.583520 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-logs" (OuterVolumeSpecName: "logs") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.594073 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b8b488c65-nmwlp"] Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.596805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6316f793-eac2-4be6-8cc1-63701fc9ca88-kube-api-access-j2qvx" (OuterVolumeSpecName: "kube-api-access-j2qvx") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "kube-api-access-j2qvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.598478 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-scripts" (OuterVolumeSpecName: "scripts") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.613403 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.630974 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.631675 4678 generic.go:334] "Generic (PLEG): container finished" podID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerID="35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3" exitCode=143 Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.631702 4678 generic.go:334] "Generic (PLEG): container finished" podID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerID="60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda" exitCode=143 Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.631750 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6316f793-eac2-4be6-8cc1-63701fc9ca88","Type":"ContainerDied","Data":"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3"} Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.631781 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6316f793-eac2-4be6-8cc1-63701fc9ca88","Type":"ContainerDied","Data":"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda"} Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.631794 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6316f793-eac2-4be6-8cc1-63701fc9ca88","Type":"ContainerDied","Data":"d6a8e5560e3e57ca1b984af0c58f0de3508feebfdcf09071b1914eb04f9313e0"} Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.631812 4678 scope.go:117] "RemoveContainer" containerID="35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.631947 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.634438 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5b97647d4d-lzbd6"] Dec 06 10:55:38 crc kubenswrapper[4678]: E1206 10:55:38.634834 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-log" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.634853 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-log" Dec 06 10:55:38 crc kubenswrapper[4678]: E1206 10:55:38.634868 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-httpd" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.634874 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-httpd" Dec 06 10:55:38 crc kubenswrapper[4678]: E1206 10:55:38.634895 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-log" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.634900 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-log" Dec 06 10:55:38 crc kubenswrapper[4678]: E1206 10:55:38.634916 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-httpd" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.634922 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-httpd" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.635099 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-httpd" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.635121 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-log" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.635129 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" containerName="glance-log" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.635148 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" containerName="glance-httpd" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.635996 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.644431 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.658699 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b97647d4d-lzbd6"] Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682138 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlmpt\" (UniqueName: \"kubernetes.io/projected/e92229b3-87c1-45dd-986b-a362109b6a83-kube-api-access-vlmpt\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682193 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-scripts\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682325 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-public-tls-certs\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682401 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-httpd-run\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682425 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682528 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-config-data\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682776 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-combined-ca-bundle\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.682836 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-logs\") pod \"e92229b3-87c1-45dd-986b-a362109b6a83\" (UID: \"e92229b3-87c1-45dd-986b-a362109b6a83\") " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683074 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-config-data\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683096 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3692143-6629-40ba-984b-187126e4a2ed-logs\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683117 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-secret-key\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683156 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzkqw\" (UniqueName: \"kubernetes.io/projected/c3692143-6629-40ba-984b-187126e4a2ed-kube-api-access-gzkqw\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683218 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-scripts\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683240 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-tls-certs\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683262 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-combined-ca-bundle\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683324 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683336 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2qvx\" (UniqueName: \"kubernetes.io/projected/6316f793-eac2-4be6-8cc1-63701fc9ca88-kube-api-access-j2qvx\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683346 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683354 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.683362 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6316f793-eac2-4be6-8cc1-63701fc9ca88-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.685322 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-logs" (OuterVolumeSpecName: "logs") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.697021 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.706672 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-scripts" (OuterVolumeSpecName: "scripts") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.711383 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.712311 4678 generic.go:334] "Generic (PLEG): container finished" podID="e92229b3-87c1-45dd-986b-a362109b6a83" containerID="aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6" exitCode=143 Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.712338 4678 generic.go:334] "Generic (PLEG): container finished" podID="e92229b3-87c1-45dd-986b-a362109b6a83" containerID="8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96" exitCode=143 Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.712359 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e92229b3-87c1-45dd-986b-a362109b6a83","Type":"ContainerDied","Data":"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6"} Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.712387 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e92229b3-87c1-45dd-986b-a362109b6a83","Type":"ContainerDied","Data":"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96"} Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.712397 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e92229b3-87c1-45dd-986b-a362109b6a83","Type":"ContainerDied","Data":"20419c2302600717e96c185bb87fe817e112f76fc9554417f4cf54fd2ff44e1f"} Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.712458 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.737463 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e92229b3-87c1-45dd-986b-a362109b6a83-kube-api-access-vlmpt" (OuterVolumeSpecName: "kube-api-access-vlmpt") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "kube-api-access-vlmpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.760362 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786015 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-config-data\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786051 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3692143-6629-40ba-984b-187126e4a2ed-logs\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786089 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-secret-key\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786126 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzkqw\" (UniqueName: \"kubernetes.io/projected/c3692143-6629-40ba-984b-187126e4a2ed-kube-api-access-gzkqw\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786201 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-scripts\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786222 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-tls-certs\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-combined-ca-bundle\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786326 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786336 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786345 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786354 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786362 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92229b3-87c1-45dd-986b-a362109b6a83-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.786370 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlmpt\" (UniqueName: \"kubernetes.io/projected/e92229b3-87c1-45dd-986b-a362109b6a83-kube-api-access-vlmpt\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.789180 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3692143-6629-40ba-984b-187126e4a2ed-logs\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.790161 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-config-data\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.790273 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-scripts\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.842354 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-dbfccf5c9-tfx4w"] Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.849785 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-config-data" (OuterVolumeSpecName: "config-data") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.849900 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.850122 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-secret-key\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.860291 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-tls-certs\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.864100 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzkqw\" (UniqueName: \"kubernetes.io/projected/c3692143-6629-40ba-984b-187126e4a2ed-kube-api-access-gzkqw\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.879220 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d77bcbbf8-cfc4l"] Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.880684 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.888508 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-combined-ca-bundle\") pod \"horizon-5b97647d4d-lzbd6\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.889766 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.889791 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.905321 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d77bcbbf8-cfc4l"] Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.918655 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6316f793-eac2-4be6-8cc1-63701fc9ca88" (UID: "6316f793-eac2-4be6-8cc1-63701fc9ca88"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.978540 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.978684 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.987694 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-config-data" (OuterVolumeSpecName: "config-data") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991610 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-combined-ca-bundle\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991671 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-horizon-secret-key\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991703 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-horizon-tls-certs\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991743 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjcg9\" (UniqueName: \"kubernetes.io/projected/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-kube-api-access-tjcg9\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991769 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-scripts\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991822 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-config-data\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991842 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-logs\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991922 4678 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6316f793-eac2-4be6-8cc1-63701fc9ca88-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991936 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.991949 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.996256 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 06 10:55:38 crc kubenswrapper[4678]: I1206 10:55:38.997571 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e92229b3-87c1-45dd-986b-a362109b6a83" (UID: "e92229b3-87c1-45dd-986b-a362109b6a83"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.013195 4678 scope.go:117] "RemoveContainer" containerID="60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.066833 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.087328 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093245 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-combined-ca-bundle\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093302 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-horizon-secret-key\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093332 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-horizon-tls-certs\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093376 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjcg9\" (UniqueName: \"kubernetes.io/projected/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-kube-api-access-tjcg9\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093395 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-scripts\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093422 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-config-data\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093435 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-logs\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093534 4678 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e92229b3-87c1-45dd-986b-a362109b6a83-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.093548 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.095747 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-logs\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.096420 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-scripts\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.097225 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-config-data\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.108380 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.113817 4678 scope.go:117] "RemoveContainer" containerID="35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.114986 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-horizon-secret-key\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: E1206 10:55:39.115221 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3\": container with ID starting with 35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3 not found: ID does not exist" containerID="35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.115305 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3"} err="failed to get container status \"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3\": rpc error: code = NotFound desc = could not find container \"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3\": container with ID starting with 35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3 not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.115375 4678 scope.go:117] "RemoveContainer" containerID="60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.115613 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-horizon-tls-certs\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: E1206 10:55:39.116053 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda\": container with ID starting with 60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda not found: ID does not exist" containerID="60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.116133 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda"} err="failed to get container status \"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda\": rpc error: code = NotFound desc = could not find container \"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda\": container with ID starting with 60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.116195 4678 scope.go:117] "RemoveContainer" containerID="35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.116451 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3"} err="failed to get container status \"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3\": rpc error: code = NotFound desc = could not find container \"35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3\": container with ID starting with 35bf7af7a9eef7a3e78f94a02ac0b7fe1b15fbae4ca4435f37435980465897c3 not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.119610 4678 scope.go:117] "RemoveContainer" containerID="60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.117130 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-combined-ca-bundle\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.120726 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda"} err="failed to get container status \"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda\": rpc error: code = NotFound desc = could not find container \"60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda\": container with ID starting with 60095d51bea99cf8b22c19889fe5420f769b40bd620dcfb9e0871173632c5fda not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.120782 4678 scope.go:117] "RemoveContainer" containerID="aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.128221 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjcg9\" (UniqueName: \"kubernetes.io/projected/c2ceb969-bd57-4345-840c-4cd8b2b7ca8e-kube-api-access-tjcg9\") pod \"horizon-7d77bcbbf8-cfc4l\" (UID: \"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e\") " pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.136531 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.137938 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.151729 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.151960 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-xrkxl" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.152131 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.151745 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.158565 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.181403 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.183409 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.194774 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.205690 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.205859 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzhwt\" (UniqueName: \"kubernetes.io/projected/d9aac132-0acd-48d8-a2e7-2d3b82157677-kube-api-access-lzhwt\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.206000 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.206026 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-logs\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.206066 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.206146 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.206195 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.206337 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.206355 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.208965 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.210669 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.296814 4678 scope.go:117] "RemoveContainer" containerID="8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.306673 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307737 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307758 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-logs\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307779 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307811 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307833 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307860 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307889 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.307933 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzhwt\" (UniqueName: \"kubernetes.io/projected/d9aac132-0acd-48d8-a2e7-2d3b82157677-kube-api-access-lzhwt\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.308777 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-logs\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.308946 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.309430 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.322477 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.325305 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.325798 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.342346 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzhwt\" (UniqueName: \"kubernetes.io/projected/d9aac132-0acd-48d8-a2e7-2d3b82157677-kube-api-access-lzhwt\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.342919 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.384057 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.410564 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-logs\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.410611 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.410691 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-config-data\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.410711 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.410737 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdvbv\" (UniqueName: \"kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.411122 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.411141 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.411162 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.444735 4678 scope.go:117] "RemoveContainer" containerID="aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6" Dec 06 10:55:39 crc kubenswrapper[4678]: E1206 10:55:39.446234 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6\": container with ID starting with aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6 not found: ID does not exist" containerID="aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.446273 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6"} err="failed to get container status \"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6\": rpc error: code = NotFound desc = could not find container \"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6\": container with ID starting with aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6 not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.446295 4678 scope.go:117] "RemoveContainer" containerID="8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96" Dec 06 10:55:39 crc kubenswrapper[4678]: E1206 10:55:39.448148 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96\": container with ID starting with 8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96 not found: ID does not exist" containerID="8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.448175 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96"} err="failed to get container status \"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96\": rpc error: code = NotFound desc = could not find container \"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96\": container with ID starting with 8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96 not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.448194 4678 scope.go:117] "RemoveContainer" containerID="aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.450416 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6"} err="failed to get container status \"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6\": rpc error: code = NotFound desc = could not find container \"aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6\": container with ID starting with aa17a66076e62c424c2aa930ae9758a8e41da7985340be758e51382481f3fed6 not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.450439 4678 scope.go:117] "RemoveContainer" containerID="8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.453446 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96"} err="failed to get container status \"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96\": rpc error: code = NotFound desc = could not find container \"8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96\": container with ID starting with 8ad0b05e0c157fa74fb95f8b27a60c4d8a6afe3d2f0caca8d293c6a79b024f96 not found: ID does not exist" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.493874 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6316f793-eac2-4be6-8cc1-63701fc9ca88" path="/var/lib/kubelet/pods/6316f793-eac2-4be6-8cc1-63701fc9ca88/volumes" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.494663 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e92229b3-87c1-45dd-986b-a362109b6a83" path="/var/lib/kubelet/pods/e92229b3-87c1-45dd-986b-a362109b6a83/volumes" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.511915 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512332 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512386 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512418 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512471 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-logs\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512516 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512629 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-config-data\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512656 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.512686 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdvbv\" (UniqueName: \"kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.513146 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-logs\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.514302 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.518159 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.518585 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.519129 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.525569 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-config-data\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.526482 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.538369 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdvbv\" (UniqueName: \"kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.542145 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " pod="openstack/glance-default-external-api-0" Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.790158 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5b97647d4d-lzbd6"] Dec 06 10:55:39 crc kubenswrapper[4678]: I1206 10:55:39.848176 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:55:40 crc kubenswrapper[4678]: I1206 10:55:40.374108 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d77bcbbf8-cfc4l"] Dec 06 10:55:40 crc kubenswrapper[4678]: I1206 10:55:40.766789 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerStarted","Data":"835d99378e15f7ba1a46a2c045e9db87f8dbdd6f5874ca7cc09d864908cbb271"} Dec 06 10:55:40 crc kubenswrapper[4678]: I1206 10:55:40.772744 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d77bcbbf8-cfc4l" event={"ID":"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e","Type":"ContainerStarted","Data":"350a0b42f35cbad540c40a1c695ed62a42a292a8ae3b3a84db785492f9ba742d"} Dec 06 10:55:40 crc kubenswrapper[4678]: I1206 10:55:40.802323 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:55:40 crc kubenswrapper[4678]: I1206 10:55:40.959983 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.047759 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-x6q8z"] Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.048328 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerName="dnsmasq-dns" containerID="cri-o://7171e508824f9159b9c4a5754f92a0faa2784942ffcfb0f12f71effb41660b8b" gracePeriod=10 Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.770751 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:55:41 crc kubenswrapper[4678]: W1206 10:55:41.792852 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccee4c53_a485_4390_b65f_c2f55f2881ad.slice/crio-77ef124e65a8722c021f47604fd0bf4b30df04bbe48dc2d14365b2c0bbaeb638 WatchSource:0}: Error finding container 77ef124e65a8722c021f47604fd0bf4b30df04bbe48dc2d14365b2c0bbaeb638: Status 404 returned error can't find the container with id 77ef124e65a8722c021f47604fd0bf4b30df04bbe48dc2d14365b2c0bbaeb638 Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.794487 4678 generic.go:334] "Generic (PLEG): container finished" podID="135a09b4-e309-47bc-8905-180c0060a639" containerID="318c3bef97a83b0cbc09c87af9eff979f07e306aa5f9bd07e6ca3aad5c757815" exitCode=0 Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.794549 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-62hxf" event={"ID":"135a09b4-e309-47bc-8905-180c0060a639","Type":"ContainerDied","Data":"318c3bef97a83b0cbc09c87af9eff979f07e306aa5f9bd07e6ca3aad5c757815"} Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.816332 4678 generic.go:334] "Generic (PLEG): container finished" podID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerID="7171e508824f9159b9c4a5754f92a0faa2784942ffcfb0f12f71effb41660b8b" exitCode=0 Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.816384 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" event={"ID":"2b1b5d12-ab97-49f9-b355-8da2966ff0bc","Type":"ContainerDied","Data":"7171e508824f9159b9c4a5754f92a0faa2784942ffcfb0f12f71effb41660b8b"} Dec 06 10:55:41 crc kubenswrapper[4678]: I1206 10:55:41.828334 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9aac132-0acd-48d8-a2e7-2d3b82157677","Type":"ContainerStarted","Data":"5ed2a46b5656e3e2ee06bda2750c2049628d2b2a1b04618bb2d8983f5838d523"} Dec 06 10:55:42 crc kubenswrapper[4678]: I1206 10:55:42.843255 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9aac132-0acd-48d8-a2e7-2d3b82157677","Type":"ContainerStarted","Data":"abe07ac3ca1ab61beacbed1af31ef4a382dcd3d073c4b53028ae90606ff58fc7"} Dec 06 10:55:42 crc kubenswrapper[4678]: I1206 10:55:42.847649 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ccee4c53-a485-4390-b65f-c2f55f2881ad","Type":"ContainerStarted","Data":"77ef124e65a8722c021f47604fd0bf4b30df04bbe48dc2d14365b2c0bbaeb638"} Dec 06 10:55:43 crc kubenswrapper[4678]: I1206 10:55:43.863159 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ccee4c53-a485-4390-b65f-c2f55f2881ad","Type":"ContainerStarted","Data":"58709e930d1cfb066f66ad4f8a5d037d014cd421a4650834889f677ea07f4b15"} Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.783802 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.799191 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807382 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-combined-ca-bundle\") pod \"135a09b4-e309-47bc-8905-180c0060a639\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807435 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-scripts\") pod \"135a09b4-e309-47bc-8905-180c0060a639\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807467 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-config-data\") pod \"135a09b4-e309-47bc-8905-180c0060a639\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807483 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-svc\") pod \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807521 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-fernet-keys\") pod \"135a09b4-e309-47bc-8905-180c0060a639\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807544 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-nb\") pod \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807608 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-credential-keys\") pod \"135a09b4-e309-47bc-8905-180c0060a639\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807630 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-sb\") pod \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807652 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcm7h\" (UniqueName: \"kubernetes.io/projected/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-kube-api-access-lcm7h\") pod \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807679 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25qcr\" (UniqueName: \"kubernetes.io/projected/135a09b4-e309-47bc-8905-180c0060a639-kube-api-access-25qcr\") pod \"135a09b4-e309-47bc-8905-180c0060a639\" (UID: \"135a09b4-e309-47bc-8905-180c0060a639\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807703 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-config\") pod \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.807721 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-swift-storage-0\") pod \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\" (UID: \"2b1b5d12-ab97-49f9-b355-8da2966ff0bc\") " Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.815455 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-kube-api-access-lcm7h" (OuterVolumeSpecName: "kube-api-access-lcm7h") pod "2b1b5d12-ab97-49f9-b355-8da2966ff0bc" (UID: "2b1b5d12-ab97-49f9-b355-8da2966ff0bc"). InnerVolumeSpecName "kube-api-access-lcm7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.822562 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-scripts" (OuterVolumeSpecName: "scripts") pod "135a09b4-e309-47bc-8905-180c0060a639" (UID: "135a09b4-e309-47bc-8905-180c0060a639"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.834442 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "135a09b4-e309-47bc-8905-180c0060a639" (UID: "135a09b4-e309-47bc-8905-180c0060a639"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.855782 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "135a09b4-e309-47bc-8905-180c0060a639" (UID: "135a09b4-e309-47bc-8905-180c0060a639"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.872370 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/135a09b4-e309-47bc-8905-180c0060a639-kube-api-access-25qcr" (OuterVolumeSpecName: "kube-api-access-25qcr") pod "135a09b4-e309-47bc-8905-180c0060a639" (UID: "135a09b4-e309-47bc-8905-180c0060a639"). InnerVolumeSpecName "kube-api-access-25qcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.896532 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-62hxf" event={"ID":"135a09b4-e309-47bc-8905-180c0060a639","Type":"ContainerDied","Data":"73ad18e37c074b820a85eb6aad94deb6d01cd99775bafa9ceae61e03e7a69ed5"} Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.896570 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73ad18e37c074b820a85eb6aad94deb6d01cd99775bafa9ceae61e03e7a69ed5" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.896629 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-62hxf" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.903375 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "135a09b4-e309-47bc-8905-180c0060a639" (UID: "135a09b4-e309-47bc-8905-180c0060a639"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.905511 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" event={"ID":"2b1b5d12-ab97-49f9-b355-8da2966ff0bc","Type":"ContainerDied","Data":"8453b948385999c8d39ffedbfcc90db5084ffc2f1a1aa879c133f32e8865dddf"} Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.905572 4678 scope.go:117] "RemoveContainer" containerID="7171e508824f9159b9c4a5754f92a0faa2784942ffcfb0f12f71effb41660b8b" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.905703 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.909242 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.909261 4678 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.909271 4678 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.909280 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcm7h\" (UniqueName: \"kubernetes.io/projected/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-kube-api-access-lcm7h\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.909289 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25qcr\" (UniqueName: \"kubernetes.io/projected/135a09b4-e309-47bc-8905-180c0060a639-kube-api-access-25qcr\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.909297 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.912230 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2b1b5d12-ab97-49f9-b355-8da2966ff0bc" (UID: "2b1b5d12-ab97-49f9-b355-8da2966ff0bc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.916845 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-config" (OuterVolumeSpecName: "config") pod "2b1b5d12-ab97-49f9-b355-8da2966ff0bc" (UID: "2b1b5d12-ab97-49f9-b355-8da2966ff0bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.929726 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-config-data" (OuterVolumeSpecName: "config-data") pod "135a09b4-e309-47bc-8905-180c0060a639" (UID: "135a09b4-e309-47bc-8905-180c0060a639"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.934249 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2b1b5d12-ab97-49f9-b355-8da2966ff0bc" (UID: "2b1b5d12-ab97-49f9-b355-8da2966ff0bc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.941035 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2b1b5d12-ab97-49f9-b355-8da2966ff0bc" (UID: "2b1b5d12-ab97-49f9-b355-8da2966ff0bc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:45 crc kubenswrapper[4678]: I1206 10:55:45.951846 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b1b5d12-ab97-49f9-b355-8da2966ff0bc" (UID: "2b1b5d12-ab97-49f9-b355-8da2966ff0bc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.010905 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.010936 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135a09b4-e309-47bc-8905-180c0060a639-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.010946 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.010955 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.010963 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.010973 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2b1b5d12-ab97-49f9-b355-8da2966ff0bc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.249949 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-x6q8z"] Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.253147 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-x6q8z"] Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.963268 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-62hxf"] Dec 06 10:55:46 crc kubenswrapper[4678]: I1206 10:55:46.982561 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-62hxf"] Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.074542 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-j8h79"] Dec 06 10:55:47 crc kubenswrapper[4678]: E1206 10:55:47.074911 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerName="dnsmasq-dns" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.074932 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerName="dnsmasq-dns" Dec 06 10:55:47 crc kubenswrapper[4678]: E1206 10:55:47.074951 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerName="init" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.074958 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerName="init" Dec 06 10:55:47 crc kubenswrapper[4678]: E1206 10:55:47.074975 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135a09b4-e309-47bc-8905-180c0060a639" containerName="keystone-bootstrap" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.074984 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="135a09b4-e309-47bc-8905-180c0060a639" containerName="keystone-bootstrap" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.075191 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="135a09b4-e309-47bc-8905-180c0060a639" containerName="keystone-bootstrap" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.075223 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerName="dnsmasq-dns" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.075978 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.079924 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.079826 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.080279 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.080328 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.080471 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z5rpk" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.101103 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-j8h79"] Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.229612 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-credential-keys\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.229685 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-scripts\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.229929 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqr5h\" (UniqueName: \"kubernetes.io/projected/6584c00a-aebd-4daf-b5db-e5cb59978e86-kube-api-access-xqr5h\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.230054 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-config-data\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.230094 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-combined-ca-bundle\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.230138 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-fernet-keys\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.331740 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-credential-keys\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.331805 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-scripts\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.331855 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqr5h\" (UniqueName: \"kubernetes.io/projected/6584c00a-aebd-4daf-b5db-e5cb59978e86-kube-api-access-xqr5h\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.331895 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-config-data\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.331915 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-combined-ca-bundle\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.331939 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-fernet-keys\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.337459 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-combined-ca-bundle\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.338021 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-fernet-keys\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.338692 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-scripts\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.347361 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-config-data\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.349865 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-credential-keys\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.350894 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqr5h\" (UniqueName: \"kubernetes.io/projected/6584c00a-aebd-4daf-b5db-e5cb59978e86-kube-api-access-xqr5h\") pod \"keystone-bootstrap-j8h79\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.409568 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.486691 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="135a09b4-e309-47bc-8905-180c0060a639" path="/var/lib/kubelet/pods/135a09b4-e309-47bc-8905-180c0060a639/volumes" Dec 06 10:55:47 crc kubenswrapper[4678]: I1206 10:55:47.487421 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" path="/var/lib/kubelet/pods/2b1b5d12-ab97-49f9-b355-8da2966ff0bc/volumes" Dec 06 10:55:49 crc kubenswrapper[4678]: I1206 10:55:49.372664 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-x6q8z" podUID="2b1b5d12-ab97-49f9-b355-8da2966ff0bc" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Dec 06 10:55:59 crc kubenswrapper[4678]: I1206 10:55:59.024102 4678 generic.go:334] "Generic (PLEG): container finished" podID="cca26a10-9756-4a74-8a33-688f36c21b46" containerID="3f0590eb23cf2fba33a719d3ba3fb99b45e39f94ba69a13180224a0a50bc72e2" exitCode=0 Dec 06 10:55:59 crc kubenswrapper[4678]: I1206 10:55:59.024165 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lmpj6" event={"ID":"cca26a10-9756-4a74-8a33-688f36c21b46","Type":"ContainerDied","Data":"3f0590eb23cf2fba33a719d3ba3fb99b45e39f94ba69a13180224a0a50bc72e2"} Dec 06 10:55:59 crc kubenswrapper[4678]: I1206 10:55:59.505661 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:55:59 crc kubenswrapper[4678]: I1206 10:55:59.505728 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:55:59 crc kubenswrapper[4678]: I1206 10:55:59.505771 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:55:59 crc kubenswrapper[4678]: I1206 10:55:59.506561 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"996baf59448dd4fd88041c7dc359f0a860305f2ace3dab71db17f1d751ccc5a8"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 10:55:59 crc kubenswrapper[4678]: I1206 10:55:59.506635 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://996baf59448dd4fd88041c7dc359f0a860305f2ace3dab71db17f1d751ccc5a8" gracePeriod=600 Dec 06 10:56:00 crc kubenswrapper[4678]: I1206 10:56:00.035655 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="996baf59448dd4fd88041c7dc359f0a860305f2ace3dab71db17f1d751ccc5a8" exitCode=0 Dec 06 10:56:00 crc kubenswrapper[4678]: I1206 10:56:00.035724 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"996baf59448dd4fd88041c7dc359f0a860305f2ace3dab71db17f1d751ccc5a8"} Dec 06 10:56:05 crc kubenswrapper[4678]: E1206 10:56:05.902171 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 06 10:56:05 crc kubenswrapper[4678]: E1206 10:56:05.904008 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5cdh5c5h5cbh5fh555h54fhf9h676h57fhd9h5b4h64chf7h56bh666hf6h56ch78h695h596hcdh66hbfh5b4h5cch89hc5h669h5f9h644h5bch5cdq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tjcg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7d77bcbbf8-cfc4l_openstack(c2ceb969-bd57-4345-840c-4cd8b2b7ca8e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:56:05 crc kubenswrapper[4678]: E1206 10:56:05.909285 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" Dec 06 10:56:06 crc kubenswrapper[4678]: E1206 10:56:06.095982 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" Dec 06 10:56:08 crc kubenswrapper[4678]: E1206 10:56:08.677939 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 06 10:56:08 crc kubenswrapper[4678]: E1206 10:56:08.678658 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n57ch8fh598h5h599hf8h56h5d7hf7hfdh99h648h66bh7ch595h5f7h89h4h65h8dh679h64fh79h5cbh656h68ch7chd4h569h676hf8hb5q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gvf7v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(3854c241-241b-49c9-a095-bac632e89d5b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:56:09 crc kubenswrapper[4678]: E1206 10:56:09.258376 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 06 10:56:09 crc kubenswrapper[4678]: E1206 10:56:09.258794 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n2hrb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-xj2n5_openstack(3a95e322-b0b3-4083-81a3-187b579c53f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:56:09 crc kubenswrapper[4678]: E1206 10:56:09.260601 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-xj2n5" podUID="3a95e322-b0b3-4083-81a3-187b579c53f4" Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.342562 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.369986 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-config\") pod \"cca26a10-9756-4a74-8a33-688f36c21b46\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.370072 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjs4j\" (UniqueName: \"kubernetes.io/projected/cca26a10-9756-4a74-8a33-688f36c21b46-kube-api-access-xjs4j\") pod \"cca26a10-9756-4a74-8a33-688f36c21b46\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.371602 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-combined-ca-bundle\") pod \"cca26a10-9756-4a74-8a33-688f36c21b46\" (UID: \"cca26a10-9756-4a74-8a33-688f36c21b46\") " Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.378021 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cca26a10-9756-4a74-8a33-688f36c21b46-kube-api-access-xjs4j" (OuterVolumeSpecName: "kube-api-access-xjs4j") pod "cca26a10-9756-4a74-8a33-688f36c21b46" (UID: "cca26a10-9756-4a74-8a33-688f36c21b46"). InnerVolumeSpecName "kube-api-access-xjs4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.398779 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-config" (OuterVolumeSpecName: "config") pod "cca26a10-9756-4a74-8a33-688f36c21b46" (UID: "cca26a10-9756-4a74-8a33-688f36c21b46"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.404901 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cca26a10-9756-4a74-8a33-688f36c21b46" (UID: "cca26a10-9756-4a74-8a33-688f36c21b46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.478678 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.478712 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/cca26a10-9756-4a74-8a33-688f36c21b46-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:09 crc kubenswrapper[4678]: I1206 10:56:09.478721 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjs4j\" (UniqueName: \"kubernetes.io/projected/cca26a10-9756-4a74-8a33-688f36c21b46-kube-api-access-xjs4j\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.126988 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lmpj6" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.127584 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lmpj6" event={"ID":"cca26a10-9756-4a74-8a33-688f36c21b46","Type":"ContainerDied","Data":"9104a188fd88f2622173de3cdebd2e46a283f659f5e2e684e7b39449117c926f"} Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.127621 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9104a188fd88f2622173de3cdebd2e46a283f659f5e2e684e7b39449117c926f" Dec 06 10:56:10 crc kubenswrapper[4678]: E1206 10:56:10.128617 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-xj2n5" podUID="3a95e322-b0b3-4083-81a3-187b579c53f4" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.560314 4678 scope.go:117] "RemoveContainer" containerID="a0deb11492465300ef7937c8d1f139355d7646e2849951ce842e5cce8ffdcd91" Dec 06 10:56:10 crc kubenswrapper[4678]: E1206 10:56:10.603633 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 06 10:56:10 crc kubenswrapper[4678]: E1206 10:56:10.603794 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q9qtz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-c448f_openstack(c315badb-6f3b-408f-b68f-b359cea62f98): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.604602 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-2f2r2"] Dec 06 10:56:10 crc kubenswrapper[4678]: E1206 10:56:10.605674 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-c448f" podUID="c315badb-6f3b-408f-b68f-b359cea62f98" Dec 06 10:56:10 crc kubenswrapper[4678]: E1206 10:56:10.609753 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cca26a10-9756-4a74-8a33-688f36c21b46" containerName="neutron-db-sync" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.609779 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cca26a10-9756-4a74-8a33-688f36c21b46" containerName="neutron-db-sync" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.610151 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cca26a10-9756-4a74-8a33-688f36c21b46" containerName="neutron-db-sync" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.611042 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.617051 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-2f2r2"] Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.715534 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-svc\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.715586 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.715663 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzz59\" (UniqueName: \"kubernetes.io/projected/0510beed-12f3-4505-9aec-cb6563d35885-kube-api-access-nzz59\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.715683 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.715716 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.715744 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-config\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.817366 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzz59\" (UniqueName: \"kubernetes.io/projected/0510beed-12f3-4505-9aec-cb6563d35885-kube-api-access-nzz59\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.817659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.817906 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.817941 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-config\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.817997 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-svc\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.818022 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.819081 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.822816 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-svc\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.822893 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.823348 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.827064 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-config\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.835685 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5f4858bb8d-jvvsq"] Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.837150 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.840222 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.840261 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8hhm2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.840423 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.840787 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.852028 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f4858bb8d-jvvsq"] Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.872908 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzz59\" (UniqueName: \"kubernetes.io/projected/0510beed-12f3-4505-9aec-cb6563d35885-kube-api-access-nzz59\") pod \"dnsmasq-dns-6b7b667979-2f2r2\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.919149 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-config\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.919233 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-ovndb-tls-certs\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.919251 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2ccb\" (UniqueName: \"kubernetes.io/projected/c939ac67-2678-47b0-9a57-7a10f89799bd-kube-api-access-h2ccb\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.919317 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-combined-ca-bundle\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:10 crc kubenswrapper[4678]: I1206 10:56:10.919583 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-httpd-config\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.023305 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-httpd-config\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.023383 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-config\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.023457 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-ovndb-tls-certs\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.023473 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2ccb\" (UniqueName: \"kubernetes.io/projected/c939ac67-2678-47b0-9a57-7a10f89799bd-kube-api-access-h2ccb\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.023559 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-combined-ca-bundle\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.028083 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-combined-ca-bundle\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.040305 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-httpd-config\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.040390 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-ovndb-tls-certs\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.043181 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-config\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.044359 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2ccb\" (UniqueName: \"kubernetes.io/projected/c939ac67-2678-47b0-9a57-7a10f89799bd-kube-api-access-h2ccb\") pod \"neutron-5f4858bb8d-jvvsq\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.045511 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:11 crc kubenswrapper[4678]: E1206 10:56:11.143049 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-c448f" podUID="c315badb-6f3b-408f-b68f-b359cea62f98" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.227639 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.229441 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-j8h79"] Dec 06 10:56:11 crc kubenswrapper[4678]: I1206 10:56:11.367946 4678 scope.go:117] "RemoveContainer" containerID="a520bd68028a84131872db9f2785f29074c8e74fb3018d3a7e8874376510222f" Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.150211 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-2f2r2"] Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.193769 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"c7aaeb71fc254443c399a6d76a377d0a6c35a168b9037ac2247e9eecdf49d005"} Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.224203 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cvhc9" event={"ID":"8fbf7713-1e99-4a84-87d0-82803fb8d0e4","Type":"ContainerStarted","Data":"4f8fd36da47d9da3f108b03408507b0bf225441036bb98d6d84c8a026018f317"} Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.247202 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6f6bc877-zxzjz" event={"ID":"66418cf7-c293-4ddc-948c-60666d9fd1c4","Type":"ContainerStarted","Data":"a0e6f9543fba06b3c541e2257f838486ec66f360a9c1271df1d580d070760d62"} Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.268670 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j8h79" event={"ID":"6584c00a-aebd-4daf-b5db-e5cb59978e86","Type":"ContainerStarted","Data":"13e34f933d799fe522f0bd65b00d3c3ef282f1e61d5d13adf24ebaaf6b67e908"} Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.268714 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j8h79" event={"ID":"6584c00a-aebd-4daf-b5db-e5cb59978e86","Type":"ContainerStarted","Data":"27dbb9336ccada6715ff314db1a6b3d70018ec7ad8c73350851a86eaedbf5ac2"} Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.278663 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f4858bb8d-jvvsq"] Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.329720 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-cvhc9" podStartSLOduration=7.046538528 podStartE2EDuration="43.329699593s" podCreationTimestamp="2025-12-06 10:55:29 +0000 UTC" firstStartedPulling="2025-12-06 10:55:32.44906436 +0000 UTC m=+1137.292495799" lastFinishedPulling="2025-12-06 10:56:08.732225425 +0000 UTC m=+1173.575656864" observedRunningTime="2025-12-06 10:56:12.297154164 +0000 UTC m=+1177.140585593" watchObservedRunningTime="2025-12-06 10:56:12.329699593 +0000 UTC m=+1177.173131032" Dec 06 10:56:12 crc kubenswrapper[4678]: I1206 10:56:12.354333 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-j8h79" podStartSLOduration=25.354311397 podStartE2EDuration="25.354311397s" podCreationTimestamp="2025-12-06 10:55:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:12.339042527 +0000 UTC m=+1177.182473966" watchObservedRunningTime="2025-12-06 10:56:12.354311397 +0000 UTC m=+1177.197742836" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.304840 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9aac132-0acd-48d8-a2e7-2d3b82157677","Type":"ContainerStarted","Data":"7c08a0e53ba728e5e8bef5ac77f4984309f75903c6d92ea13330a8497fd15033"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.332217 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6f6bc877-zxzjz" event={"ID":"66418cf7-c293-4ddc-948c-60666d9fd1c4","Type":"ContainerStarted","Data":"f20d4dcd8b390c979698d3d4e59545b67f8a42853778bd20da961122a56f8010"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.332389 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c6f6bc877-zxzjz" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon-log" containerID="cri-o://a0e6f9543fba06b3c541e2257f838486ec66f360a9c1271df1d580d070760d62" gracePeriod=30 Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.333004 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c6f6bc877-zxzjz" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon" containerID="cri-o://f20d4dcd8b390c979698d3d4e59545b67f8a42853778bd20da961122a56f8010" gracePeriod=30 Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.333128 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=34.333106651 podStartE2EDuration="34.333106651s" podCreationTimestamp="2025-12-06 10:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:13.326359153 +0000 UTC m=+1178.169790592" watchObservedRunningTime="2025-12-06 10:56:13.333106651 +0000 UTC m=+1178.176538090" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.355534 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ccee4c53-a485-4390-b65f-c2f55f2881ad","Type":"ContainerStarted","Data":"6200b7dd3e61dde1e396ff21451edde208c22b656f3b16792eddc1031dc13ecb"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.360780 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c6f6bc877-zxzjz" podStartSLOduration=6.266568614 podStartE2EDuration="44.360769255s" podCreationTimestamp="2025-12-06 10:55:29 +0000 UTC" firstStartedPulling="2025-12-06 10:55:31.170971187 +0000 UTC m=+1136.014402626" lastFinishedPulling="2025-12-06 10:56:09.265171808 +0000 UTC m=+1174.108603267" observedRunningTime="2025-12-06 10:56:13.358697194 +0000 UTC m=+1178.202128633" watchObservedRunningTime="2025-12-06 10:56:13.360769255 +0000 UTC m=+1178.204200694" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.361296 4678 generic.go:334] "Generic (PLEG): container finished" podID="0510beed-12f3-4505-9aec-cb6563d35885" containerID="b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae" exitCode=0 Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.361389 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" event={"ID":"0510beed-12f3-4505-9aec-cb6563d35885","Type":"ContainerDied","Data":"b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.361419 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" event={"ID":"0510beed-12f3-4505-9aec-cb6563d35885","Type":"ContainerStarted","Data":"1021d77d737c50ee54b270dbd1e418ec9b992485f039039bbb753e3e97d8e100"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.366245 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerStarted","Data":"fd37d0f403224fe0b9ae5fdd56a3b0dd389c3b987d4dfe06f0a769ded18b7265"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.370447 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4858bb8d-jvvsq" event={"ID":"c939ac67-2678-47b0-9a57-7a10f89799bd","Type":"ContainerStarted","Data":"a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.370476 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4858bb8d-jvvsq" event={"ID":"c939ac67-2678-47b0-9a57-7a10f89799bd","Type":"ContainerStarted","Data":"1133de2a92c0dbc1b7e9131bea8061275eb1d3fd468d0939e8e44e3dbfa668c5"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.381428 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b488c65-nmwlp" event={"ID":"13c9e9db-8842-4412-9d09-4f8fdb8a39cf","Type":"ContainerStarted","Data":"5a73f5df43ec07225878944207d903870895424133964a5124b8e351950f50fa"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.381716 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8b488c65-nmwlp" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon-log" containerID="cri-o://5a73f5df43ec07225878944207d903870895424133964a5124b8e351950f50fa" gracePeriod=30 Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.381947 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8b488c65-nmwlp" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon" containerID="cri-o://0c6771cf0d5d25a0f33802927088ac645141b77ff4375191cdcd0be341faeea3" gracePeriod=30 Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.384502 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerStarted","Data":"389817da78af88665e5f2000c353529adec4c3df99e02f2713ce867a856272b2"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.384539 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerStarted","Data":"d4163c241eb326f5ee5663d607080cbbfd1967af2e9381c39c27f1ceaeebb861"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.407030 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=34.406976695 podStartE2EDuration="34.406976695s" podCreationTimestamp="2025-12-06 10:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:13.398952949 +0000 UTC m=+1178.242384388" watchObservedRunningTime="2025-12-06 10:56:13.406976695 +0000 UTC m=+1178.250408134" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.429586 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-dbfccf5c9-tfx4w" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon-log" containerID="cri-o://1cc46eeca1eaffbcb92eb5d37e993d0510583804e6a803421637bf4b0a388830" gracePeriod=30 Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.429892 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbfccf5c9-tfx4w" event={"ID":"6faea971-f662-49d1-b48c-5743f127d7b5","Type":"ContainerStarted","Data":"3623d2ee1b98fd054f950002a6caaa770eb1c83a5b39063c09a7899ec4d7a50c"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.429919 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbfccf5c9-tfx4w" event={"ID":"6faea971-f662-49d1-b48c-5743f127d7b5","Type":"ContainerStarted","Data":"1cc46eeca1eaffbcb92eb5d37e993d0510583804e6a803421637bf4b0a388830"} Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.431013 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-dbfccf5c9-tfx4w" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon" containerID="cri-o://3623d2ee1b98fd054f950002a6caaa770eb1c83a5b39063c09a7899ec4d7a50c" gracePeriod=30 Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.458950 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b8b488c65-nmwlp" podStartSLOduration=5.241909258 podStartE2EDuration="43.458926783s" podCreationTimestamp="2025-12-06 10:55:30 +0000 UTC" firstStartedPulling="2025-12-06 10:55:33.154742407 +0000 UTC m=+1137.998173846" lastFinishedPulling="2025-12-06 10:56:11.371759932 +0000 UTC m=+1176.215191371" observedRunningTime="2025-12-06 10:56:13.441569342 +0000 UTC m=+1178.285000781" watchObservedRunningTime="2025-12-06 10:56:13.458926783 +0000 UTC m=+1178.302358232" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.513225 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5b97647d4d-lzbd6" podStartSLOduration=4.15840665 podStartE2EDuration="35.513206671s" podCreationTimestamp="2025-12-06 10:55:38 +0000 UTC" firstStartedPulling="2025-12-06 10:55:39.959986314 +0000 UTC m=+1144.803417753" lastFinishedPulling="2025-12-06 10:56:11.314786335 +0000 UTC m=+1176.158217774" observedRunningTime="2025-12-06 10:56:13.503496785 +0000 UTC m=+1178.346928224" watchObservedRunningTime="2025-12-06 10:56:13.513206671 +0000 UTC m=+1178.356638110" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.554740 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-dbfccf5c9-tfx4w" podStartSLOduration=4.018607594 podStartE2EDuration="40.554716502s" podCreationTimestamp="2025-12-06 10:55:33 +0000 UTC" firstStartedPulling="2025-12-06 10:55:34.889717734 +0000 UTC m=+1139.733149173" lastFinishedPulling="2025-12-06 10:56:11.425826642 +0000 UTC m=+1176.269258081" observedRunningTime="2025-12-06 10:56:13.534855528 +0000 UTC m=+1178.378286967" watchObservedRunningTime="2025-12-06 10:56:13.554716502 +0000 UTC m=+1178.398147951" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.928455 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7567449455-km9cw"] Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.930421 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.934116 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.934304 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 06 10:56:13 crc kubenswrapper[4678]: I1206 10:56:13.956074 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7567449455-km9cw"] Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.020824 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.107977 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-httpd-config\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.108043 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-internal-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.108072 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-ovndb-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.108111 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-config\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.108151 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-public-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.108174 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-combined-ca-bundle\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.108194 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blgfb\" (UniqueName: \"kubernetes.io/projected/0e386ecc-df37-4e08-a9f9-80abb7019c19-kube-api-access-blgfb\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.209964 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-httpd-config\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.210198 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-internal-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.210298 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-ovndb-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.210421 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-config\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.210584 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-public-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.210671 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-combined-ca-bundle\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.210739 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blgfb\" (UniqueName: \"kubernetes.io/projected/0e386ecc-df37-4e08-a9f9-80abb7019c19-kube-api-access-blgfb\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.222553 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-httpd-config\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.224004 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-internal-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.239117 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-public-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.239725 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-ovndb-tls-certs\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.241033 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-combined-ca-bundle\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.248833 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blgfb\" (UniqueName: \"kubernetes.io/projected/0e386ecc-df37-4e08-a9f9-80abb7019c19-kube-api-access-blgfb\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.249647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/0e386ecc-df37-4e08-a9f9-80abb7019c19-config\") pod \"neutron-7567449455-km9cw\" (UID: \"0e386ecc-df37-4e08-a9f9-80abb7019c19\") " pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.265379 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.440564 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4858bb8d-jvvsq" event={"ID":"c939ac67-2678-47b0-9a57-7a10f89799bd","Type":"ContainerStarted","Data":"7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae"} Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.441670 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.449463 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b488c65-nmwlp" event={"ID":"13c9e9db-8842-4412-9d09-4f8fdb8a39cf","Type":"ContainerStarted","Data":"0c6771cf0d5d25a0f33802927088ac645141b77ff4375191cdcd0be341faeea3"} Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.454257 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" event={"ID":"0510beed-12f3-4505-9aec-cb6563d35885","Type":"ContainerStarted","Data":"956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc"} Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.454295 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.468811 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5f4858bb8d-jvvsq" podStartSLOduration=4.468793343 podStartE2EDuration="4.468793343s" podCreationTimestamp="2025-12-06 10:56:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:14.467770282 +0000 UTC m=+1179.311201721" watchObservedRunningTime="2025-12-06 10:56:14.468793343 +0000 UTC m=+1179.312224782" Dec 06 10:56:14 crc kubenswrapper[4678]: I1206 10:56:14.491791 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" podStartSLOduration=4.491771468 podStartE2EDuration="4.491771468s" podCreationTimestamp="2025-12-06 10:56:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:14.488002338 +0000 UTC m=+1179.331433777" watchObservedRunningTime="2025-12-06 10:56:14.491771468 +0000 UTC m=+1179.335202907" Dec 06 10:56:15 crc kubenswrapper[4678]: I1206 10:56:15.113225 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7567449455-km9cw"] Dec 06 10:56:15 crc kubenswrapper[4678]: W1206 10:56:15.130568 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e386ecc_df37_4e08_a9f9_80abb7019c19.slice/crio-5d4b94848e719efda04de501a901d671607bcff38b570f173fd70c61afe7e4d5 WatchSource:0}: Error finding container 5d4b94848e719efda04de501a901d671607bcff38b570f173fd70c61afe7e4d5: Status 404 returned error can't find the container with id 5d4b94848e719efda04de501a901d671607bcff38b570f173fd70c61afe7e4d5 Dec 06 10:56:15 crc kubenswrapper[4678]: I1206 10:56:15.464753 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7567449455-km9cw" event={"ID":"0e386ecc-df37-4e08-a9f9-80abb7019c19","Type":"ContainerStarted","Data":"5d4b94848e719efda04de501a901d671607bcff38b570f173fd70c61afe7e4d5"} Dec 06 10:56:18 crc kubenswrapper[4678]: I1206 10:56:18.510027 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7567449455-km9cw" event={"ID":"0e386ecc-df37-4e08-a9f9-80abb7019c19","Type":"ContainerStarted","Data":"319bdd673973fb4115dbb8e03d6c4c34107d2de45674acae564978f8f45d604a"} Dec 06 10:56:18 crc kubenswrapper[4678]: I1206 10:56:18.976702 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:56:18 crc kubenswrapper[4678]: I1206 10:56:18.976749 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.512463 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.512540 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.519650 4678 generic.go:334] "Generic (PLEG): container finished" podID="8fbf7713-1e99-4a84-87d0-82803fb8d0e4" containerID="4f8fd36da47d9da3f108b03408507b0bf225441036bb98d6d84c8a026018f317" exitCode=0 Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.519690 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cvhc9" event={"ID":"8fbf7713-1e99-4a84-87d0-82803fb8d0e4","Type":"ContainerDied","Data":"4f8fd36da47d9da3f108b03408507b0bf225441036bb98d6d84c8a026018f317"} Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.576037 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.576113 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.850017 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.850329 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.906188 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.908308 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 10:56:19 crc kubenswrapper[4678]: I1206 10:56:19.963741 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:56:20 crc kubenswrapper[4678]: I1206 10:56:20.529188 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:20 crc kubenswrapper[4678]: I1206 10:56:20.529247 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 10:56:20 crc kubenswrapper[4678]: I1206 10:56:20.529259 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 10:56:20 crc kubenswrapper[4678]: I1206 10:56:20.529422 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:20 crc kubenswrapper[4678]: I1206 10:56:20.909113 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:56:21 crc kubenswrapper[4678]: I1206 10:56:21.047679 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:21 crc kubenswrapper[4678]: I1206 10:56:21.115584 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-tfqd9"] Dec 06 10:56:21 crc kubenswrapper[4678]: I1206 10:56:21.115892 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" podUID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerName="dnsmasq-dns" containerID="cri-o://1b5d79aec9cd09780fce89be3ee6ffb3ee057a47e1e10ac244770d45a54a903d" gracePeriod=10 Dec 06 10:56:21 crc kubenswrapper[4678]: I1206 10:56:21.538312 4678 generic.go:334] "Generic (PLEG): container finished" podID="6584c00a-aebd-4daf-b5db-e5cb59978e86" containerID="13e34f933d799fe522f0bd65b00d3c3ef282f1e61d5d13adf24ebaaf6b67e908" exitCode=0 Dec 06 10:56:21 crc kubenswrapper[4678]: I1206 10:56:21.538476 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j8h79" event={"ID":"6584c00a-aebd-4daf-b5db-e5cb59978e86","Type":"ContainerDied","Data":"13e34f933d799fe522f0bd65b00d3c3ef282f1e61d5d13adf24ebaaf6b67e908"} Dec 06 10:56:21 crc kubenswrapper[4678]: I1206 10:56:21.541385 4678 generic.go:334] "Generic (PLEG): container finished" podID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerID="1b5d79aec9cd09780fce89be3ee6ffb3ee057a47e1e10ac244770d45a54a903d" exitCode=0 Dec 06 10:56:21 crc kubenswrapper[4678]: I1206 10:56:21.541970 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" event={"ID":"ced0ddf7-759e-4cb9-9516-02174759fddf","Type":"ContainerDied","Data":"1b5d79aec9cd09780fce89be3ee6ffb3ee057a47e1e10ac244770d45a54a903d"} Dec 06 10:56:22 crc kubenswrapper[4678]: I1206 10:56:22.973210 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cvhc9" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.002054 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.121975 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-fernet-keys\") pod \"6584c00a-aebd-4daf-b5db-e5cb59978e86\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122033 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-logs\") pod \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122060 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-combined-ca-bundle\") pod \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122081 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-config-data\") pod \"6584c00a-aebd-4daf-b5db-e5cb59978e86\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122131 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-combined-ca-bundle\") pod \"6584c00a-aebd-4daf-b5db-e5cb59978e86\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122154 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqr5h\" (UniqueName: \"kubernetes.io/projected/6584c00a-aebd-4daf-b5db-e5cb59978e86-kube-api-access-xqr5h\") pod \"6584c00a-aebd-4daf-b5db-e5cb59978e86\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122238 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-config-data\") pod \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122279 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-credential-keys\") pod \"6584c00a-aebd-4daf-b5db-e5cb59978e86\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122307 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k456h\" (UniqueName: \"kubernetes.io/projected/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-kube-api-access-k456h\") pod \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122411 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-scripts\") pod \"6584c00a-aebd-4daf-b5db-e5cb59978e86\" (UID: \"6584c00a-aebd-4daf-b5db-e5cb59978e86\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.122441 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-scripts\") pod \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\" (UID: \"8fbf7713-1e99-4a84-87d0-82803fb8d0e4\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.126833 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-logs" (OuterVolumeSpecName: "logs") pod "8fbf7713-1e99-4a84-87d0-82803fb8d0e4" (UID: "8fbf7713-1e99-4a84-87d0-82803fb8d0e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.130090 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-scripts" (OuterVolumeSpecName: "scripts") pod "8fbf7713-1e99-4a84-87d0-82803fb8d0e4" (UID: "8fbf7713-1e99-4a84-87d0-82803fb8d0e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.136993 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-kube-api-access-k456h" (OuterVolumeSpecName: "kube-api-access-k456h") pod "8fbf7713-1e99-4a84-87d0-82803fb8d0e4" (UID: "8fbf7713-1e99-4a84-87d0-82803fb8d0e4"). InnerVolumeSpecName "kube-api-access-k456h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.151707 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6584c00a-aebd-4daf-b5db-e5cb59978e86-kube-api-access-xqr5h" (OuterVolumeSpecName: "kube-api-access-xqr5h") pod "6584c00a-aebd-4daf-b5db-e5cb59978e86" (UID: "6584c00a-aebd-4daf-b5db-e5cb59978e86"). InnerVolumeSpecName "kube-api-access-xqr5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.153042 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6584c00a-aebd-4daf-b5db-e5cb59978e86" (UID: "6584c00a-aebd-4daf-b5db-e5cb59978e86"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.161263 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6584c00a-aebd-4daf-b5db-e5cb59978e86" (UID: "6584c00a-aebd-4daf-b5db-e5cb59978e86"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.161382 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-scripts" (OuterVolumeSpecName: "scripts") pod "6584c00a-aebd-4daf-b5db-e5cb59978e86" (UID: "6584c00a-aebd-4daf-b5db-e5cb59978e86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.189844 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-config-data" (OuterVolumeSpecName: "config-data") pod "6584c00a-aebd-4daf-b5db-e5cb59978e86" (UID: "6584c00a-aebd-4daf-b5db-e5cb59978e86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.194319 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-config-data" (OuterVolumeSpecName: "config-data") pod "8fbf7713-1e99-4a84-87d0-82803fb8d0e4" (UID: "8fbf7713-1e99-4a84-87d0-82803fb8d0e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.202334 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fbf7713-1e99-4a84-87d0-82803fb8d0e4" (UID: "8fbf7713-1e99-4a84-87d0-82803fb8d0e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.211829 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6584c00a-aebd-4daf-b5db-e5cb59978e86" (UID: "6584c00a-aebd-4daf-b5db-e5cb59978e86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224088 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224121 4678 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224134 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k456h\" (UniqueName: \"kubernetes.io/projected/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-kube-api-access-k456h\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224146 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224153 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224161 4678 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224169 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224178 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbf7713-1e99-4a84-87d0-82803fb8d0e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224187 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224195 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6584c00a-aebd-4daf-b5db-e5cb59978e86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.224203 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqr5h\" (UniqueName: \"kubernetes.io/projected/6584c00a-aebd-4daf-b5db-e5cb59978e86-kube-api-access-xqr5h\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.610151 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cvhc9" event={"ID":"8fbf7713-1e99-4a84-87d0-82803fb8d0e4","Type":"ContainerDied","Data":"84108bcfa8b30b7cf8f509396c6e6218e64d9877765413ceffc107810fc3c875"} Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.610516 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84108bcfa8b30b7cf8f509396c6e6218e64d9877765413ceffc107810fc3c875" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.610587 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cvhc9" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.626647 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d77bcbbf8-cfc4l" event={"ID":"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e","Type":"ContainerStarted","Data":"9e9c073cc4eb7000b8ed4b9c51608f8ecd98c7c6420f0535469d89885da6bd47"} Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.643718 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-j8h79" event={"ID":"6584c00a-aebd-4daf-b5db-e5cb59978e86","Type":"ContainerDied","Data":"27dbb9336ccada6715ff314db1a6b3d70018ec7ad8c73350851a86eaedbf5ac2"} Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.643907 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27dbb9336ccada6715ff314db1a6b3d70018ec7ad8c73350851a86eaedbf5ac2" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.644034 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-j8h79" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.750271 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7f69d54bd9-rd99t"] Dec 06 10:56:23 crc kubenswrapper[4678]: E1206 10:56:23.750986 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbf7713-1e99-4a84-87d0-82803fb8d0e4" containerName="placement-db-sync" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.751000 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbf7713-1e99-4a84-87d0-82803fb8d0e4" containerName="placement-db-sync" Dec 06 10:56:23 crc kubenswrapper[4678]: E1206 10:56:23.751025 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6584c00a-aebd-4daf-b5db-e5cb59978e86" containerName="keystone-bootstrap" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.751031 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6584c00a-aebd-4daf-b5db-e5cb59978e86" containerName="keystone-bootstrap" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.751646 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fbf7713-1e99-4a84-87d0-82803fb8d0e4" containerName="placement-db-sync" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.751659 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6584c00a-aebd-4daf-b5db-e5cb59978e86" containerName="keystone-bootstrap" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.774555 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.781332 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.781691 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.781939 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-z5rpk" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.782154 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.782348 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.784250 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.793540 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f69d54bd9-rd99t"] Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845342 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-scripts\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845420 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dwbz\" (UniqueName: \"kubernetes.io/projected/eb641828-16dd-4437-8266-e20a7a27ad1c-kube-api-access-5dwbz\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845441 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-config-data\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845532 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-fernet-keys\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845579 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-public-tls-certs\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845601 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-combined-ca-bundle\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845624 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-internal-tls-certs\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.845653 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-credential-keys\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.878591 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.949754 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-nb\") pod \"ced0ddf7-759e-4cb9-9516-02174759fddf\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.949800 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-config\") pod \"ced0ddf7-759e-4cb9-9516-02174759fddf\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.949821 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-sb\") pod \"ced0ddf7-759e-4cb9-9516-02174759fddf\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.949845 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-swift-storage-0\") pod \"ced0ddf7-759e-4cb9-9516-02174759fddf\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.949878 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kj6pm\" (UniqueName: \"kubernetes.io/projected/ced0ddf7-759e-4cb9-9516-02174759fddf-kube-api-access-kj6pm\") pod \"ced0ddf7-759e-4cb9-9516-02174759fddf\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.949964 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-svc\") pod \"ced0ddf7-759e-4cb9-9516-02174759fddf\" (UID: \"ced0ddf7-759e-4cb9-9516-02174759fddf\") " Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950235 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-scripts\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950289 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dwbz\" (UniqueName: \"kubernetes.io/projected/eb641828-16dd-4437-8266-e20a7a27ad1c-kube-api-access-5dwbz\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950308 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-config-data\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950356 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-fernet-keys\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950389 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-public-tls-certs\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950408 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-combined-ca-bundle\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950428 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-internal-tls-certs\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.950453 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-credential-keys\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:23 crc kubenswrapper[4678]: I1206 10:56:23.989091 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-fernet-keys\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.009573 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-credential-keys\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.010059 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-public-tls-certs\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.023355 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dwbz\" (UniqueName: \"kubernetes.io/projected/eb641828-16dd-4437-8266-e20a7a27ad1c-kube-api-access-5dwbz\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.032837 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-scripts\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.033368 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-combined-ca-bundle\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.067601 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ced0ddf7-759e-4cb9-9516-02174759fddf-kube-api-access-kj6pm" (OuterVolumeSpecName: "kube-api-access-kj6pm") pod "ced0ddf7-759e-4cb9-9516-02174759fddf" (UID: "ced0ddf7-759e-4cb9-9516-02174759fddf"). InnerVolumeSpecName "kube-api-access-kj6pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.069721 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-config-data\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.070079 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kj6pm\" (UniqueName: \"kubernetes.io/projected/ced0ddf7-759e-4cb9-9516-02174759fddf-kube-api-access-kj6pm\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.070426 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb641828-16dd-4437-8266-e20a7a27ad1c-internal-tls-certs\") pod \"keystone-7f69d54bd9-rd99t\" (UID: \"eb641828-16dd-4437-8266-e20a7a27ad1c\") " pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.173649 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.236393 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5bd747c594-kkr56"] Dec 06 10:56:24 crc kubenswrapper[4678]: E1206 10:56:24.236771 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerName="dnsmasq-dns" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.236785 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerName="dnsmasq-dns" Dec 06 10:56:24 crc kubenswrapper[4678]: E1206 10:56:24.236799 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerName="init" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.236805 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerName="init" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.236976 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ced0ddf7-759e-4cb9-9516-02174759fddf" containerName="dnsmasq-dns" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.237961 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5bd747c594-kkr56"] Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.238050 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.242763 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.242955 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.243098 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.246117 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-pr6nm" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.246083 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.287229 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ced0ddf7-759e-4cb9-9516-02174759fddf" (UID: "ced0ddf7-759e-4cb9-9516-02174759fddf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.288659 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.333122 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ced0ddf7-759e-4cb9-9516-02174759fddf" (UID: "ced0ddf7-759e-4cb9-9516-02174759fddf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.337379 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ced0ddf7-759e-4cb9-9516-02174759fddf" (UID: "ced0ddf7-759e-4cb9-9516-02174759fddf"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.340994 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ced0ddf7-759e-4cb9-9516-02174759fddf" (UID: "ced0ddf7-759e-4cb9-9516-02174759fddf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.388217 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-config" (OuterVolumeSpecName: "config") pod "ced0ddf7-759e-4cb9-9516-02174759fddf" (UID: "ced0ddf7-759e-4cb9-9516-02174759fddf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.394846 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-internal-tls-certs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.394885 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-public-tls-certs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.394932 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e209518e-5689-468e-9b70-9c1e2864eb9b-logs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.394961 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qcjp\" (UniqueName: \"kubernetes.io/projected/e209518e-5689-468e-9b70-9c1e2864eb9b-kube-api-access-5qcjp\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.394985 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-combined-ca-bundle\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.395022 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-scripts\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.395048 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-config-data\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.395123 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.395135 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.395143 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.395154 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ced0ddf7-759e-4cb9-9516-02174759fddf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.497089 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-scripts\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.497149 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-config-data\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.497221 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-internal-tls-certs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.497237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-public-tls-certs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.497281 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e209518e-5689-468e-9b70-9c1e2864eb9b-logs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.497310 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qcjp\" (UniqueName: \"kubernetes.io/projected/e209518e-5689-468e-9b70-9c1e2864eb9b-kube-api-access-5qcjp\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.497331 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-combined-ca-bundle\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.500997 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e209518e-5689-468e-9b70-9c1e2864eb9b-logs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.509508 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-combined-ca-bundle\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.509730 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-scripts\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.510154 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-internal-tls-certs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.516459 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-public-tls-certs\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.517360 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e209518e-5689-468e-9b70-9c1e2864eb9b-config-data\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.533098 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qcjp\" (UniqueName: \"kubernetes.io/projected/e209518e-5689-468e-9b70-9c1e2864eb9b-kube-api-access-5qcjp\") pod \"placement-5bd747c594-kkr56\" (UID: \"e209518e-5689-468e-9b70-9c1e2864eb9b\") " pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.579882 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.753504 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7f69d54bd9-rd99t"] Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.763036 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d77bcbbf8-cfc4l" event={"ID":"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e","Type":"ContainerStarted","Data":"a04f288a96a3e3ff2987759b3e148e1cbbffe2e84ab32bb6357b33ca748488f8"} Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.803778 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d77bcbbf8-cfc4l" podStartSLOduration=-9223371990.051012 podStartE2EDuration="46.803763354s" podCreationTimestamp="2025-12-06 10:55:38 +0000 UTC" firstStartedPulling="2025-12-06 10:55:40.400325863 +0000 UTC m=+1145.243757302" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:24.80362716 +0000 UTC m=+1189.647058599" watchObservedRunningTime="2025-12-06 10:56:24.803763354 +0000 UTC m=+1189.647194793" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.807518 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerStarted","Data":"d7cb112822079b5ada3b60642d202ec2cb12e813e32e4c0b22bdc86a8f2c076c"} Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.837750 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" event={"ID":"ced0ddf7-759e-4cb9-9516-02174759fddf","Type":"ContainerDied","Data":"eca2934a21b03f93da8b56dcebf47aeb801a9e13dcb0cbdd3c397ac04aaeb611"} Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.837830 4678 scope.go:117] "RemoveContainer" containerID="1b5d79aec9cd09780fce89be3ee6ffb3ee057a47e1e10ac244770d45a54a903d" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.837963 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-tfqd9" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.860652 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7567449455-km9cw" event={"ID":"0e386ecc-df37-4e08-a9f9-80abb7019c19","Type":"ContainerStarted","Data":"d068e1140af9d92b834a348c12c96288707e2a62fd54e41dc5876511f1764e5d"} Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.861763 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.904665 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7567449455-km9cw" podStartSLOduration=11.904636392 podStartE2EDuration="11.904636392s" podCreationTimestamp="2025-12-06 10:56:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:24.886294962 +0000 UTC m=+1189.729726401" watchObservedRunningTime="2025-12-06 10:56:24.904636392 +0000 UTC m=+1189.748067841" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.942981 4678 scope.go:117] "RemoveContainer" containerID="64d4ce6c62f1f0b60b8cccb02adf9ff56c28776a70da86031032edf2db0f5ac9" Dec 06 10:56:24 crc kubenswrapper[4678]: I1206 10:56:24.989538 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-tfqd9"] Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.012254 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-tfqd9"] Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.507820 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ced0ddf7-759e-4cb9-9516-02174759fddf" path="/var/lib/kubelet/pods/ced0ddf7-759e-4cb9-9516-02174759fddf/volumes" Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.539645 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5bd747c594-kkr56"] Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.877017 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5bd747c594-kkr56" event={"ID":"e209518e-5689-468e-9b70-9c1e2864eb9b","Type":"ContainerStarted","Data":"7402a6588bd6436513e2ca8f09087060e65316afbe81ad4e982150b579272102"} Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.890857 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f69d54bd9-rd99t" event={"ID":"eb641828-16dd-4437-8266-e20a7a27ad1c","Type":"ContainerStarted","Data":"24041668f242f7a2d47d0a08b321595b79b24e76557d247d4ac6d06f9d631796"} Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.890894 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7f69d54bd9-rd99t" event={"ID":"eb641828-16dd-4437-8266-e20a7a27ad1c","Type":"ContainerStarted","Data":"97ed556527d6263244deb8cbedbe63b945e8e6636dabc2656f2088e746c143a5"} Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.891550 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:25 crc kubenswrapper[4678]: I1206 10:56:25.931989 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7f69d54bd9-rd99t" podStartSLOduration=2.931971635 podStartE2EDuration="2.931971635s" podCreationTimestamp="2025-12-06 10:56:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:25.926074622 +0000 UTC m=+1190.769506061" watchObservedRunningTime="2025-12-06 10:56:25.931971635 +0000 UTC m=+1190.775403064" Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.914183 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj2n5" event={"ID":"3a95e322-b0b3-4083-81a3-187b579c53f4","Type":"ContainerStarted","Data":"66ef8a822ff594231a59cc28f0a007e97e0d17aa5883855c56365e8ee7c84ee2"} Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.920607 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c448f" event={"ID":"c315badb-6f3b-408f-b68f-b359cea62f98","Type":"ContainerStarted","Data":"a5649c76e4a471f9fd4a47f9a86959923c310d68810204b448fc42b47cab289c"} Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.933381 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5bd747c594-kkr56" event={"ID":"e209518e-5689-468e-9b70-9c1e2864eb9b","Type":"ContainerStarted","Data":"35cb5611514588382f572bd819a4b643e5633595d73cadd0de8ed5fd26c125c1"} Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.933420 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5bd747c594-kkr56" event={"ID":"e209518e-5689-468e-9b70-9c1e2864eb9b","Type":"ContainerStarted","Data":"799be6b115491fb44e7f0ae7cded0e322e6cd191dc1f4d6a6098fdf8f6c29fe6"} Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.933713 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.933765 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.950189 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xj2n5" podStartSLOduration=3.853760057 podStartE2EDuration="56.950169979s" podCreationTimestamp="2025-12-06 10:55:30 +0000 UTC" firstStartedPulling="2025-12-06 10:55:33.325842932 +0000 UTC m=+1138.169274371" lastFinishedPulling="2025-12-06 10:56:26.422252854 +0000 UTC m=+1191.265684293" observedRunningTime="2025-12-06 10:56:26.938063463 +0000 UTC m=+1191.781494902" watchObservedRunningTime="2025-12-06 10:56:26.950169979 +0000 UTC m=+1191.793601418" Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.968125 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-c448f" podStartSLOduration=7.120959768 podStartE2EDuration="57.968104327s" podCreationTimestamp="2025-12-06 10:55:29 +0000 UTC" firstStartedPulling="2025-12-06 10:55:33.191454507 +0000 UTC m=+1138.034885946" lastFinishedPulling="2025-12-06 10:56:24.038599066 +0000 UTC m=+1188.882030505" observedRunningTime="2025-12-06 10:56:26.963740458 +0000 UTC m=+1191.807171897" watchObservedRunningTime="2025-12-06 10:56:26.968104327 +0000 UTC m=+1191.811535766" Dec 06 10:56:26 crc kubenswrapper[4678]: I1206 10:56:26.990370 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5bd747c594-kkr56" podStartSLOduration=2.990347072 podStartE2EDuration="2.990347072s" podCreationTimestamp="2025-12-06 10:56:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:26.983939273 +0000 UTC m=+1191.827370712" watchObservedRunningTime="2025-12-06 10:56:26.990347072 +0000 UTC m=+1191.833778511" Dec 06 10:56:27 crc kubenswrapper[4678]: I1206 10:56:27.593885 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 10:56:27 crc kubenswrapper[4678]: I1206 10:56:27.594008 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:56:27 crc kubenswrapper[4678]: I1206 10:56:27.598356 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 10:56:27 crc kubenswrapper[4678]: I1206 10:56:27.618189 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:27 crc kubenswrapper[4678]: I1206 10:56:27.618292 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:56:28 crc kubenswrapper[4678]: I1206 10:56:28.250683 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 10:56:28 crc kubenswrapper[4678]: I1206 10:56:28.978694 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:56:29 crc kubenswrapper[4678]: I1206 10:56:29.309296 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:56:29 crc kubenswrapper[4678]: I1206 10:56:29.309360 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:56:36 crc kubenswrapper[4678]: I1206 10:56:36.055406 4678 generic.go:334] "Generic (PLEG): container finished" podID="3a95e322-b0b3-4083-81a3-187b579c53f4" containerID="66ef8a822ff594231a59cc28f0a007e97e0d17aa5883855c56365e8ee7c84ee2" exitCode=0 Dec 06 10:56:36 crc kubenswrapper[4678]: I1206 10:56:36.056065 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj2n5" event={"ID":"3a95e322-b0b3-4083-81a3-187b579c53f4","Type":"ContainerDied","Data":"66ef8a822ff594231a59cc28f0a007e97e0d17aa5883855c56365e8ee7c84ee2"} Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.093221 4678 generic.go:334] "Generic (PLEG): container finished" podID="c315badb-6f3b-408f-b68f-b359cea62f98" containerID="a5649c76e4a471f9fd4a47f9a86959923c310d68810204b448fc42b47cab289c" exitCode=0 Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.093277 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c448f" event={"ID":"c315badb-6f3b-408f-b68f-b359cea62f98","Type":"ContainerDied","Data":"a5649c76e4a471f9fd4a47f9a86959923c310d68810204b448fc42b47cab289c"} Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.371125 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.442701 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2hrb\" (UniqueName: \"kubernetes.io/projected/3a95e322-b0b3-4083-81a3-187b579c53f4-kube-api-access-n2hrb\") pod \"3a95e322-b0b3-4083-81a3-187b579c53f4\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.442760 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-db-sync-config-data\") pod \"3a95e322-b0b3-4083-81a3-187b579c53f4\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.442782 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-combined-ca-bundle\") pod \"3a95e322-b0b3-4083-81a3-187b579c53f4\" (UID: \"3a95e322-b0b3-4083-81a3-187b579c53f4\") " Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.456642 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3a95e322-b0b3-4083-81a3-187b579c53f4" (UID: "3a95e322-b0b3-4083-81a3-187b579c53f4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.466617 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a95e322-b0b3-4083-81a3-187b579c53f4-kube-api-access-n2hrb" (OuterVolumeSpecName: "kube-api-access-n2hrb") pod "3a95e322-b0b3-4083-81a3-187b579c53f4" (UID: "3a95e322-b0b3-4083-81a3-187b579c53f4"). InnerVolumeSpecName "kube-api-access-n2hrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.481596 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a95e322-b0b3-4083-81a3-187b579c53f4" (UID: "3a95e322-b0b3-4083-81a3-187b579c53f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.544497 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2hrb\" (UniqueName: \"kubernetes.io/projected/3a95e322-b0b3-4083-81a3-187b579c53f4-kube-api-access-n2hrb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.544557 4678 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.544569 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a95e322-b0b3-4083-81a3-187b579c53f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:38 crc kubenswrapper[4678]: I1206 10:56:38.977037 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.105103 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xj2n5" event={"ID":"3a95e322-b0b3-4083-81a3-187b579c53f4","Type":"ContainerDied","Data":"aa5671680bcb47d6bb36e6c6eb9c4f9f4dd5940602c8bfeb4154accfd85c0181"} Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.105688 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa5671680bcb47d6bb36e6c6eb9c4f9f4dd5940602c8bfeb4154accfd85c0181" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.105127 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xj2n5" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.321691 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.624066 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c448f" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695001 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-scripts\") pod \"c315badb-6f3b-408f-b68f-b359cea62f98\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695067 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c315badb-6f3b-408f-b68f-b359cea62f98-etc-machine-id\") pod \"c315badb-6f3b-408f-b68f-b359cea62f98\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695099 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-db-sync-config-data\") pod \"c315badb-6f3b-408f-b68f-b359cea62f98\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695128 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9qtz\" (UniqueName: \"kubernetes.io/projected/c315badb-6f3b-408f-b68f-b359cea62f98-kube-api-access-q9qtz\") pod \"c315badb-6f3b-408f-b68f-b359cea62f98\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695145 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-config-data\") pod \"c315badb-6f3b-408f-b68f-b359cea62f98\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695185 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-combined-ca-bundle\") pod \"c315badb-6f3b-408f-b68f-b359cea62f98\" (UID: \"c315badb-6f3b-408f-b68f-b359cea62f98\") " Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695235 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c315badb-6f3b-408f-b68f-b359cea62f98-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c315badb-6f3b-408f-b68f-b359cea62f98" (UID: "c315badb-6f3b-408f-b68f-b359cea62f98"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.695599 4678 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c315badb-6f3b-408f-b68f-b359cea62f98-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.708884 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c315badb-6f3b-408f-b68f-b359cea62f98-kube-api-access-q9qtz" (OuterVolumeSpecName: "kube-api-access-q9qtz") pod "c315badb-6f3b-408f-b68f-b359cea62f98" (UID: "c315badb-6f3b-408f-b68f-b359cea62f98"). InnerVolumeSpecName "kube-api-access-q9qtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.713070 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-scripts" (OuterVolumeSpecName: "scripts") pod "c315badb-6f3b-408f-b68f-b359cea62f98" (UID: "c315badb-6f3b-408f-b68f-b359cea62f98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.714397 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c315badb-6f3b-408f-b68f-b359cea62f98" (UID: "c315badb-6f3b-408f-b68f-b359cea62f98"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.785209 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-config-data" (OuterVolumeSpecName: "config-data") pod "c315badb-6f3b-408f-b68f-b359cea62f98" (UID: "c315badb-6f3b-408f-b68f-b359cea62f98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.789692 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c315badb-6f3b-408f-b68f-b359cea62f98" (UID: "c315badb-6f3b-408f-b68f-b359cea62f98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.796749 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.796782 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.796792 4678 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.796801 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9qtz\" (UniqueName: \"kubernetes.io/projected/c315badb-6f3b-408f-b68f-b359cea62f98-kube-api-access-q9qtz\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.796812 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c315badb-6f3b-408f-b68f-b359cea62f98-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.842117 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5674df646-s4z6g"] Dec 06 10:56:39 crc kubenswrapper[4678]: E1206 10:56:39.842532 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a95e322-b0b3-4083-81a3-187b579c53f4" containerName="barbican-db-sync" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.842548 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a95e322-b0b3-4083-81a3-187b579c53f4" containerName="barbican-db-sync" Dec 06 10:56:39 crc kubenswrapper[4678]: E1206 10:56:39.842564 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c315badb-6f3b-408f-b68f-b359cea62f98" containerName="cinder-db-sync" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.842570 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c315badb-6f3b-408f-b68f-b359cea62f98" containerName="cinder-db-sync" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.842736 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c315badb-6f3b-408f-b68f-b359cea62f98" containerName="cinder-db-sync" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.842763 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a95e322-b0b3-4083-81a3-187b579c53f4" containerName="barbican-db-sync" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.850727 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:39 crc kubenswrapper[4678]: E1206 10:56:39.873403 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="3854c241-241b-49c9-a095-bac632e89d5b" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.896084 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.902273 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-7kv78" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.902482 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.903603 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-combined-ca-bundle\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.903664 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-logs\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.903690 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-config-data-custom\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.903720 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s856r\" (UniqueName: \"kubernetes.io/projected/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-kube-api-access-s856r\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.903773 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-config-data\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.917930 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5674df646-s4z6g"] Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.955545 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-597bfcdc9c-vdzb7"] Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.957029 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.965346 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 06 10:56:39 crc kubenswrapper[4678]: I1206 10:56:39.973745 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-597bfcdc9c-vdzb7"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.006691 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-config-data\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.006943 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-logs\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007048 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-config-data\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007130 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpvfm\" (UniqueName: \"kubernetes.io/projected/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-kube-api-access-cpvfm\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007226 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-combined-ca-bundle\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007321 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-combined-ca-bundle\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007396 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-config-data-custom\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007477 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-logs\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007630 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-config-data-custom\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007708 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s856r\" (UniqueName: \"kubernetes.io/projected/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-kube-api-access-s856r\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.007923 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-logs\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.017406 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-combined-ca-bundle\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.017616 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-config-data-custom\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.018396 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-config-data\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.060217 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s856r\" (UniqueName: \"kubernetes.io/projected/e55dbe47-fb3b-4020-bba4-f8c038e7cc5b-kube-api-access-s856r\") pod \"barbican-keystone-listener-5674df646-s4z6g\" (UID: \"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b\") " pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.108947 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-config-data\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.109011 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-logs\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.109040 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpvfm\" (UniqueName: \"kubernetes.io/projected/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-kube-api-access-cpvfm\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.109100 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-combined-ca-bundle\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.109122 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-config-data-custom\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.109582 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-logs\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.120229 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-config-data\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.121105 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-combined-ca-bundle\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.125751 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-c448f" event={"ID":"c315badb-6f3b-408f-b68f-b359cea62f98","Type":"ContainerDied","Data":"9902cb8efe854befd74db740f5f547002aeffb49ced787373b8dd46975519977"} Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.125792 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9902cb8efe854befd74db740f5f547002aeffb49ced787373b8dd46975519977" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.125858 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-c448f" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.129276 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-config-data-custom\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.148095 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpvfm\" (UniqueName: \"kubernetes.io/projected/69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957-kube-api-access-cpvfm\") pod \"barbican-worker-597bfcdc9c-vdzb7\" (UID: \"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957\") " pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.150995 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerStarted","Data":"512972e6f15d7024be99291c48b0a98b7989b536416addb0c092de8879cb1098"} Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.151134 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="ceilometer-notification-agent" containerID="cri-o://fd37d0f403224fe0b9ae5fdd56a3b0dd389c3b987d4dfe06f0a769ded18b7265" gracePeriod=30 Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.151267 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.151566 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="sg-core" containerID="cri-o://d7cb112822079b5ada3b60642d202ec2cb12e813e32e4c0b22bdc86a8f2c076c" gracePeriod=30 Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.151669 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="proxy-httpd" containerID="cri-o://512972e6f15d7024be99291c48b0a98b7989b536416addb0c092de8879cb1098" gracePeriod=30 Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.182180 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5674df646-s4z6g" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.267624 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6b4ff"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.279851 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-597bfcdc9c-vdzb7" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.280356 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.313316 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6b4ff"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.367208 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6d978978c4-qsthb"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.368909 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.387123 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.422339 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.422586 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnffq\" (UniqueName: \"kubernetes.io/projected/de91d1f3-4bc4-4611-ad00-169a5711fd54-kube-api-access-jnffq\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.422712 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd23d8eb-f06c-48e4-9319-aa699a260ca0-logs\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.422819 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-combined-ca-bundle\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.422938 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.423026 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data-custom\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.423134 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.423242 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.423338 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgvfz\" (UniqueName: \"kubernetes.io/projected/cd23d8eb-f06c-48e4-9319-aa699a260ca0-kube-api-access-jgvfz\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.423407 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.423499 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.423687 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d978978c4-qsthb"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546538 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-combined-ca-bundle\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546785 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546810 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data-custom\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546826 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546858 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546907 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgvfz\" (UniqueName: \"kubernetes.io/projected/cd23d8eb-f06c-48e4-9319-aa699a260ca0-kube-api-access-jgvfz\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546921 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.546942 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.547003 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.547021 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnffq\" (UniqueName: \"kubernetes.io/projected/de91d1f3-4bc4-4611-ad00-169a5711fd54-kube-api-access-jnffq\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.547087 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd23d8eb-f06c-48e4-9319-aa699a260ca0-logs\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.547459 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd23d8eb-f06c-48e4-9319-aa699a260ca0-logs\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.548747 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.560653 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.560813 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-combined-ca-bundle\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.561519 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.561881 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.562288 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.569200 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data-custom\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.577879 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.585343 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgvfz\" (UniqueName: \"kubernetes.io/projected/cd23d8eb-f06c-48e4-9319-aa699a260ca0-kube-api-access-jgvfz\") pod \"barbican-api-6d978978c4-qsthb\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.612713 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnffq\" (UniqueName: \"kubernetes.io/projected/de91d1f3-4bc4-4611-ad00-169a5711fd54-kube-api-access-jnffq\") pod \"dnsmasq-dns-848cf88cfc-6b4ff\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.625014 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.710848 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.714398 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.715919 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.721064 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.721277 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-g5rcq" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.721471 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.721509 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.760179 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.760259 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.760302 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.760344 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9p64\" (UniqueName: \"kubernetes.io/projected/cd48a1a0-2fe4-423e-8930-99d5347e218e-kube-api-access-c9p64\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.760368 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.760465 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd48a1a0-2fe4-423e-8930-99d5347e218e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.761836 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.862459 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6b4ff"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.863581 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.863786 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.863923 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.870331 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-scripts\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.870597 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9p64\" (UniqueName: \"kubernetes.io/projected/cd48a1a0-2fe4-423e-8930-99d5347e218e-kube-api-access-c9p64\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.870696 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.871021 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd48a1a0-2fe4-423e-8930-99d5347e218e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.871359 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd48a1a0-2fe4-423e-8930-99d5347e218e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.892976 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.895262 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.896185 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.913607 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9p64\" (UniqueName: \"kubernetes.io/projected/cd48a1a0-2fe4-423e-8930-99d5347e218e-kube-api-access-c9p64\") pod \"cinder-scheduler-0\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " pod="openstack/cinder-scheduler-0" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.948111 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-mkvr4"] Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.949649 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:40 crc kubenswrapper[4678]: I1206 10:56:40.974452 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-mkvr4"] Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.046299 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.090881 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.090951 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-config\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.091010 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn9dh\" (UniqueName: \"kubernetes.io/projected/eb1e07f0-3034-47d2-9143-82058df6d73c-kube-api-access-wn9dh\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.091056 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-svc\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.091160 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.091215 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.194337 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.195305 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.196542 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-config\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.197109 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-config\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.199015 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn9dh\" (UniqueName: \"kubernetes.io/projected/eb1e07f0-3034-47d2-9143-82058df6d73c-kube-api-access-wn9dh\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.199436 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-svc\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.200435 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-svc\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.200712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.201287 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.203719 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.203170 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.255820 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn9dh\" (UniqueName: \"kubernetes.io/projected/eb1e07f0-3034-47d2-9143-82058df6d73c-kube-api-access-wn9dh\") pod \"dnsmasq-dns-6578955fd5-mkvr4\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.271122 4678 generic.go:334] "Generic (PLEG): container finished" podID="3854c241-241b-49c9-a095-bac632e89d5b" containerID="d7cb112822079b5ada3b60642d202ec2cb12e813e32e4c0b22bdc86a8f2c076c" exitCode=2 Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.271205 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerDied","Data":"d7cb112822079b5ada3b60642d202ec2cb12e813e32e4c0b22bdc86a8f2c076c"} Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.287210 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.325234 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.327279 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.336628 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.344951 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.417395 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.454436 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-logs\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.454513 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.454655 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-scripts\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.454675 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92g7g\" (UniqueName: \"kubernetes.io/projected/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-kube-api-access-92g7g\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.454921 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.454995 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.455029 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data-custom\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.556609 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-scripts\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.556655 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92g7g\" (UniqueName: \"kubernetes.io/projected/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-kube-api-access-92g7g\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.556719 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.556755 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.556771 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data-custom\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.556822 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-logs\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.556841 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.568003 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-scripts\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.569101 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-logs\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.574114 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.585369 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.585857 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data-custom\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.618774 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.619658 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92g7g\" (UniqueName: \"kubernetes.io/projected/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-kube-api-access-92g7g\") pod \"cinder-api-0\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.620149 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-597bfcdc9c-vdzb7"] Dec 06 10:56:41 crc kubenswrapper[4678]: W1206 10:56:41.661246 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69b9a5eb_f2d2_423e_b3c6_1f4b5b8ef957.slice/crio-70e9933e19102adbeb024fac0ebf9ddf0a1c57371612b0cf29bafd671c2a2610 WatchSource:0}: Error finding container 70e9933e19102adbeb024fac0ebf9ddf0a1c57371612b0cf29bafd671c2a2610: Status 404 returned error can't find the container with id 70e9933e19102adbeb024fac0ebf9ddf0a1c57371612b0cf29bafd671c2a2610 Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.707519 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5674df646-s4z6g"] Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.713583 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.909766 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6b4ff"] Dec 06 10:56:41 crc kubenswrapper[4678]: I1206 10:56:41.930317 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6d978978c4-qsthb"] Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.043640 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.213786 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-mkvr4"] Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.304162 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" event={"ID":"de91d1f3-4bc4-4611-ad00-169a5711fd54","Type":"ContainerStarted","Data":"a923cc1c06271f6a480007eb66120f73849c489a2e3d2cb868dce37341738070"} Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.311687 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd48a1a0-2fe4-423e-8930-99d5347e218e","Type":"ContainerStarted","Data":"49a437635f70a0c63809e6624a73dab596f471dc4071070910c9fa234afb35f2"} Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.335257 4678 generic.go:334] "Generic (PLEG): container finished" podID="3854c241-241b-49c9-a095-bac632e89d5b" containerID="512972e6f15d7024be99291c48b0a98b7989b536416addb0c092de8879cb1098" exitCode=0 Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.335411 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerDied","Data":"512972e6f15d7024be99291c48b0a98b7989b536416addb0c092de8879cb1098"} Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.355615 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5674df646-s4z6g" event={"ID":"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b","Type":"ContainerStarted","Data":"5743bb6651ce849512454a8a62d6efc20e8ce8c2b641fb00fe692041a3d16080"} Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.380787 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.392544 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-597bfcdc9c-vdzb7" event={"ID":"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957","Type":"ContainerStarted","Data":"70e9933e19102adbeb024fac0ebf9ddf0a1c57371612b0cf29bafd671c2a2610"} Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.414454 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" event={"ID":"eb1e07f0-3034-47d2-9143-82058df6d73c","Type":"ContainerStarted","Data":"396de4321be4ab465449d73aaf0b8a4bd09893ea04b23095a64ce28dc90a4cf4"} Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.426050 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d978978c4-qsthb" event={"ID":"cd23d8eb-f06c-48e4-9319-aa699a260ca0","Type":"ContainerStarted","Data":"b0facf2a6703c559ab4d6fd9a8fe36d4168c234d46da2104728642e1cc47fff1"} Dec 06 10:56:42 crc kubenswrapper[4678]: I1206 10:56:42.426104 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d978978c4-qsthb" event={"ID":"cd23d8eb-f06c-48e4-9319-aa699a260ca0","Type":"ContainerStarted","Data":"f0ddfe9b686bcb88e8e88100258841c73ac8aa405c4a7cb0162eafa7ab6a5c5f"} Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.450626 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d4ece9-a8c0-4842-a3d3-d5c05c77676b","Type":"ContainerStarted","Data":"d1422f9ae758ed0af0770176be16d4db093954bb2b3d91868e8274c427f39159"} Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.488947 4678 generic.go:334] "Generic (PLEG): container finished" podID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerID="299e677dad807c4dd6e0139b8af97619a409d6baa0b114364570545bcd8ed161" exitCode=0 Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.494359 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" event={"ID":"eb1e07f0-3034-47d2-9143-82058df6d73c","Type":"ContainerDied","Data":"299e677dad807c4dd6e0139b8af97619a409d6baa0b114364570545bcd8ed161"} Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.562449 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d978978c4-qsthb" event={"ID":"cd23d8eb-f06c-48e4-9319-aa699a260ca0","Type":"ContainerStarted","Data":"0964ceddf233314aea34efa77bab3f1761d93ba02156121900794bb4df1ede45"} Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.562623 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.562699 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.578535 4678 generic.go:334] "Generic (PLEG): container finished" podID="de91d1f3-4bc4-4611-ad00-169a5711fd54" containerID="851aed5373d9231d48409ea09697ed4a29f26087951d6f0886a754c3249d8a00" exitCode=0 Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.578584 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" event={"ID":"de91d1f3-4bc4-4611-ad00-169a5711fd54","Type":"ContainerDied","Data":"851aed5373d9231d48409ea09697ed4a29f26087951d6f0886a754c3249d8a00"} Dec 06 10:56:43 crc kubenswrapper[4678]: I1206 10:56:43.606320 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6d978978c4-qsthb" podStartSLOduration=3.606305993 podStartE2EDuration="3.606305993s" podCreationTimestamp="2025-12-06 10:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:43.605255432 +0000 UTC m=+1208.448686871" watchObservedRunningTime="2025-12-06 10:56:43.606305993 +0000 UTC m=+1208.449737432" Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.291444 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7567449455-km9cw" Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.354064 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5f4858bb8d-jvvsq"] Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.354570 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5f4858bb8d-jvvsq" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-api" containerID="cri-o://a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712" gracePeriod=30 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.355145 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5f4858bb8d-jvvsq" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-httpd" containerID="cri-o://7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae" gracePeriod=30 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.590693 4678 generic.go:334] "Generic (PLEG): container finished" podID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerID="0c6771cf0d5d25a0f33802927088ac645141b77ff4375191cdcd0be341faeea3" exitCode=137 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.590720 4678 generic.go:334] "Generic (PLEG): container finished" podID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerID="5a73f5df43ec07225878944207d903870895424133964a5124b8e351950f50fa" exitCode=137 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.590777 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b488c65-nmwlp" event={"ID":"13c9e9db-8842-4412-9d09-4f8fdb8a39cf","Type":"ContainerDied","Data":"0c6771cf0d5d25a0f33802927088ac645141b77ff4375191cdcd0be341faeea3"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.590827 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b488c65-nmwlp" event={"ID":"13c9e9db-8842-4412-9d09-4f8fdb8a39cf","Type":"ContainerDied","Data":"5a73f5df43ec07225878944207d903870895424133964a5124b8e351950f50fa"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.597388 4678 generic.go:334] "Generic (PLEG): container finished" podID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerID="f20d4dcd8b390c979698d3d4e59545b67f8a42853778bd20da961122a56f8010" exitCode=137 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.597425 4678 generic.go:334] "Generic (PLEG): container finished" podID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerID="a0e6f9543fba06b3c541e2257f838486ec66f360a9c1271df1d580d070760d62" exitCode=137 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.597497 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6f6bc877-zxzjz" event={"ID":"66418cf7-c293-4ddc-948c-60666d9fd1c4","Type":"ContainerDied","Data":"f20d4dcd8b390c979698d3d4e59545b67f8a42853778bd20da961122a56f8010"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.597527 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6f6bc877-zxzjz" event={"ID":"66418cf7-c293-4ddc-948c-60666d9fd1c4","Type":"ContainerDied","Data":"a0e6f9543fba06b3c541e2257f838486ec66f360a9c1271df1d580d070760d62"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.612565 4678 generic.go:334] "Generic (PLEG): container finished" podID="6faea971-f662-49d1-b48c-5743f127d7b5" containerID="3623d2ee1b98fd054f950002a6caaa770eb1c83a5b39063c09a7899ec4d7a50c" exitCode=137 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.612604 4678 generic.go:334] "Generic (PLEG): container finished" podID="6faea971-f662-49d1-b48c-5743f127d7b5" containerID="1cc46eeca1eaffbcb92eb5d37e993d0510583804e6a803421637bf4b0a388830" exitCode=137 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.612692 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbfccf5c9-tfx4w" event={"ID":"6faea971-f662-49d1-b48c-5743f127d7b5","Type":"ContainerDied","Data":"3623d2ee1b98fd054f950002a6caaa770eb1c83a5b39063c09a7899ec4d7a50c"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.612746 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbfccf5c9-tfx4w" event={"ID":"6faea971-f662-49d1-b48c-5743f127d7b5","Type":"ContainerDied","Data":"1cc46eeca1eaffbcb92eb5d37e993d0510583804e6a803421637bf4b0a388830"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.614278 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d4ece9-a8c0-4842-a3d3-d5c05c77676b","Type":"ContainerStarted","Data":"1cc7475c58e76a748f418936dc0f0af1b64deac6832d1a31e3f736f45e4011cc"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.615857 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd48a1a0-2fe4-423e-8930-99d5347e218e","Type":"ContainerStarted","Data":"7c2b7044c50faf48f6df34973d69856dbf7569ad12ced58571ac9ff8f41badb6"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.618624 4678 generic.go:334] "Generic (PLEG): container finished" podID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerID="7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae" exitCode=0 Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.619733 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4858bb8d-jvvsq" event={"ID":"c939ac67-2678-47b0-9a57-7a10f89799bd","Type":"ContainerDied","Data":"7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae"} Dec 06 10:56:44 crc kubenswrapper[4678]: I1206 10:56:44.932657 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.383389 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.469510 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-nb\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.469566 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.469689 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.469729 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnffq\" (UniqueName: \"kubernetes.io/projected/de91d1f3-4bc4-4611-ad00-169a5711fd54-kube-api-access-jnffq\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.469822 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-svc\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.469842 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-swift-storage-0\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.525517 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de91d1f3-4bc4-4611-ad00-169a5711fd54-kube-api-access-jnffq" (OuterVolumeSpecName: "kube-api-access-jnffq") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "kube-api-access-jnffq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.553895 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.558149 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.570702 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.572109 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config" (OuterVolumeSpecName: "config") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.572629 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.572774 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config\") pod \"de91d1f3-4bc4-4611-ad00-169a5711fd54\" (UID: \"de91d1f3-4bc4-4611-ad00-169a5711fd54\") " Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.573594 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.573613 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.573623 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnffq\" (UniqueName: \"kubernetes.io/projected/de91d1f3-4bc4-4611-ad00-169a5711fd54-kube-api-access-jnffq\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:45 crc kubenswrapper[4678]: W1206 10:56:45.574516 4678 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/de91d1f3-4bc4-4611-ad00-169a5711fd54/volumes/kubernetes.io~configmap/ovsdbserver-sb Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.574533 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: W1206 10:56:45.574573 4678 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/de91d1f3-4bc4-4611-ad00-169a5711fd54/volumes/kubernetes.io~configmap/config Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.574580 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config" (OuterVolumeSpecName: "config") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.620098 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "de91d1f3-4bc4-4611-ad00-169a5711fd54" (UID: "de91d1f3-4bc4-4611-ad00-169a5711fd54"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.674292 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.674330 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.674339 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de91d1f3-4bc4-4611-ad00-169a5711fd54-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.675168 4678 generic.go:334] "Generic (PLEG): container finished" podID="3854c241-241b-49c9-a095-bac632e89d5b" containerID="fd37d0f403224fe0b9ae5fdd56a3b0dd389c3b987d4dfe06f0a769ded18b7265" exitCode=0 Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.675219 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerDied","Data":"fd37d0f403224fe0b9ae5fdd56a3b0dd389c3b987d4dfe06f0a769ded18b7265"} Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.712060 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" event={"ID":"de91d1f3-4bc4-4611-ad00-169a5711fd54","Type":"ContainerDied","Data":"a923cc1c06271f6a480007eb66120f73849c489a2e3d2cb868dce37341738070"} Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.712112 4678 scope.go:117] "RemoveContainer" containerID="851aed5373d9231d48409ea09697ed4a29f26087951d6f0886a754c3249d8a00" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.712241 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-6b4ff" Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.829623 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6b4ff"] Dec 06 10:56:45 crc kubenswrapper[4678]: I1206 10:56:45.836722 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-6b4ff"] Dec 06 10:56:46 crc kubenswrapper[4678]: I1206 10:56:46.726311 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8b488c65-nmwlp" event={"ID":"13c9e9db-8842-4412-9d09-4f8fdb8a39cf","Type":"ContainerDied","Data":"8f2c5a23b1403337b829e5345838fd2dfe9bfa318b04e4a83f38ec53a40278e5"} Dec 06 10:56:46 crc kubenswrapper[4678]: I1206 10:56:46.726592 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f2c5a23b1403337b829e5345838fd2dfe9bfa318b04e4a83f38ec53a40278e5" Dec 06 10:56:46 crc kubenswrapper[4678]: I1206 10:56:46.730562 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c6f6bc877-zxzjz" event={"ID":"66418cf7-c293-4ddc-948c-60666d9fd1c4","Type":"ContainerDied","Data":"fd8f5f6c7062b21b224751e56d6b035c9f50ec25d55c1bdb6d5a7fb2a1472599"} Dec 06 10:56:46 crc kubenswrapper[4678]: I1206 10:56:46.730717 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd8f5f6c7062b21b224751e56d6b035c9f50ec25d55c1bdb6d5a7fb2a1472599" Dec 06 10:56:46 crc kubenswrapper[4678]: I1206 10:56:46.854105 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:56:46 crc kubenswrapper[4678]: I1206 10:56:46.893954 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:56:46 crc kubenswrapper[4678]: I1206 10:56:46.894846 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008479 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-config-data\") pod \"66418cf7-c293-4ddc-948c-60666d9fd1c4\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008528 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gmdh\" (UniqueName: \"kubernetes.io/projected/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-kube-api-access-2gmdh\") pod \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008546 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-scripts\") pod \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008571 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-logs\") pod \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008587 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcw8b\" (UniqueName: \"kubernetes.io/projected/6faea971-f662-49d1-b48c-5743f127d7b5-kube-api-access-fcw8b\") pod \"6faea971-f662-49d1-b48c-5743f127d7b5\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008600 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6faea971-f662-49d1-b48c-5743f127d7b5-logs\") pod \"6faea971-f662-49d1-b48c-5743f127d7b5\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008617 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-scripts\") pod \"6faea971-f662-49d1-b48c-5743f127d7b5\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008640 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-config-data\") pod \"6faea971-f662-49d1-b48c-5743f127d7b5\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008695 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-scripts\") pod \"66418cf7-c293-4ddc-948c-60666d9fd1c4\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008714 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66418cf7-c293-4ddc-948c-60666d9fd1c4-horizon-secret-key\") pod \"66418cf7-c293-4ddc-948c-60666d9fd1c4\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008742 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-horizon-secret-key\") pod \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008757 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-config-data\") pod \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\" (UID: \"13c9e9db-8842-4412-9d09-4f8fdb8a39cf\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008800 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwchh\" (UniqueName: \"kubernetes.io/projected/66418cf7-c293-4ddc-948c-60666d9fd1c4-kube-api-access-rwchh\") pod \"66418cf7-c293-4ddc-948c-60666d9fd1c4\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008826 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6faea971-f662-49d1-b48c-5743f127d7b5-horizon-secret-key\") pod \"6faea971-f662-49d1-b48c-5743f127d7b5\" (UID: \"6faea971-f662-49d1-b48c-5743f127d7b5\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.008842 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66418cf7-c293-4ddc-948c-60666d9fd1c4-logs\") pod \"66418cf7-c293-4ddc-948c-60666d9fd1c4\" (UID: \"66418cf7-c293-4ddc-948c-60666d9fd1c4\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.009460 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66418cf7-c293-4ddc-948c-60666d9fd1c4-logs" (OuterVolumeSpecName: "logs") pod "66418cf7-c293-4ddc-948c-60666d9fd1c4" (UID: "66418cf7-c293-4ddc-948c-60666d9fd1c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.011130 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-logs" (OuterVolumeSpecName: "logs") pod "13c9e9db-8842-4412-9d09-4f8fdb8a39cf" (UID: "13c9e9db-8842-4412-9d09-4f8fdb8a39cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.017169 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6faea971-f662-49d1-b48c-5743f127d7b5-logs" (OuterVolumeSpecName: "logs") pod "6faea971-f662-49d1-b48c-5743f127d7b5" (UID: "6faea971-f662-49d1-b48c-5743f127d7b5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.022326 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-kube-api-access-2gmdh" (OuterVolumeSpecName: "kube-api-access-2gmdh") pod "13c9e9db-8842-4412-9d09-4f8fdb8a39cf" (UID: "13c9e9db-8842-4412-9d09-4f8fdb8a39cf"). InnerVolumeSpecName "kube-api-access-2gmdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.022896 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "13c9e9db-8842-4412-9d09-4f8fdb8a39cf" (UID: "13c9e9db-8842-4412-9d09-4f8fdb8a39cf"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.067725 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66418cf7-c293-4ddc-948c-60666d9fd1c4-kube-api-access-rwchh" (OuterVolumeSpecName: "kube-api-access-rwchh") pod "66418cf7-c293-4ddc-948c-60666d9fd1c4" (UID: "66418cf7-c293-4ddc-948c-60666d9fd1c4"). InnerVolumeSpecName "kube-api-access-rwchh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.073240 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6faea971-f662-49d1-b48c-5743f127d7b5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "6faea971-f662-49d1-b48c-5743f127d7b5" (UID: "6faea971-f662-49d1-b48c-5743f127d7b5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.082235 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6faea971-f662-49d1-b48c-5743f127d7b5-kube-api-access-fcw8b" (OuterVolumeSpecName: "kube-api-access-fcw8b") pod "6faea971-f662-49d1-b48c-5743f127d7b5" (UID: "6faea971-f662-49d1-b48c-5743f127d7b5"). InnerVolumeSpecName "kube-api-access-fcw8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.097055 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66418cf7-c293-4ddc-948c-60666d9fd1c4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "66418cf7-c293-4ddc-948c-60666d9fd1c4" (UID: "66418cf7-c293-4ddc-948c-60666d9fd1c4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114623 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6faea971-f662-49d1-b48c-5743f127d7b5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114658 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66418cf7-c293-4ddc-948c-60666d9fd1c4-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114668 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gmdh\" (UniqueName: \"kubernetes.io/projected/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-kube-api-access-2gmdh\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114680 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114690 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcw8b\" (UniqueName: \"kubernetes.io/projected/6faea971-f662-49d1-b48c-5743f127d7b5-kube-api-access-fcw8b\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114698 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6faea971-f662-49d1-b48c-5743f127d7b5-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114707 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/66418cf7-c293-4ddc-948c-60666d9fd1c4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114716 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.114724 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwchh\" (UniqueName: \"kubernetes.io/projected/66418cf7-c293-4ddc-948c-60666d9fd1c4-kube-api-access-rwchh\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.191399 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-config-data" (OuterVolumeSpecName: "config-data") pod "13c9e9db-8842-4412-9d09-4f8fdb8a39cf" (UID: "13c9e9db-8842-4412-9d09-4f8fdb8a39cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.200076 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-config-data" (OuterVolumeSpecName: "config-data") pod "6faea971-f662-49d1-b48c-5743f127d7b5" (UID: "6faea971-f662-49d1-b48c-5743f127d7b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.211394 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-config-data" (OuterVolumeSpecName: "config-data") pod "66418cf7-c293-4ddc-948c-60666d9fd1c4" (UID: "66418cf7-c293-4ddc-948c-60666d9fd1c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.211587 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-scripts" (OuterVolumeSpecName: "scripts") pod "13c9e9db-8842-4412-9d09-4f8fdb8a39cf" (UID: "13c9e9db-8842-4412-9d09-4f8fdb8a39cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.225667 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.225693 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.225704 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/13c9e9db-8842-4412-9d09-4f8fdb8a39cf-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.225713 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.280069 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-scripts" (OuterVolumeSpecName: "scripts") pod "6faea971-f662-49d1-b48c-5743f127d7b5" (UID: "6faea971-f662-49d1-b48c-5743f127d7b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.282955 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-scripts" (OuterVolumeSpecName: "scripts") pod "66418cf7-c293-4ddc-948c-60666d9fd1c4" (UID: "66418cf7-c293-4ddc-948c-60666d9fd1c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.326233 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/66418cf7-c293-4ddc-948c-60666d9fd1c4-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.326257 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6faea971-f662-49d1-b48c-5743f127d7b5-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.445552 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.510599 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de91d1f3-4bc4-4611-ad00-169a5711fd54" path="/var/lib/kubelet/pods/de91d1f3-4bc4-4611-ad00-169a5711fd54/volumes" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.631132 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-log-httpd\") pod \"3854c241-241b-49c9-a095-bac632e89d5b\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.631306 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-scripts\") pod \"3854c241-241b-49c9-a095-bac632e89d5b\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.631337 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvf7v\" (UniqueName: \"kubernetes.io/projected/3854c241-241b-49c9-a095-bac632e89d5b-kube-api-access-gvf7v\") pod \"3854c241-241b-49c9-a095-bac632e89d5b\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.631398 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-sg-core-conf-yaml\") pod \"3854c241-241b-49c9-a095-bac632e89d5b\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.631516 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-config-data\") pod \"3854c241-241b-49c9-a095-bac632e89d5b\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.631552 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-run-httpd\") pod \"3854c241-241b-49c9-a095-bac632e89d5b\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.631576 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-combined-ca-bundle\") pod \"3854c241-241b-49c9-a095-bac632e89d5b\" (UID: \"3854c241-241b-49c9-a095-bac632e89d5b\") " Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.633848 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3854c241-241b-49c9-a095-bac632e89d5b" (UID: "3854c241-241b-49c9-a095-bac632e89d5b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.635048 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3854c241-241b-49c9-a095-bac632e89d5b" (UID: "3854c241-241b-49c9-a095-bac632e89d5b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.645129 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3854c241-241b-49c9-a095-bac632e89d5b-kube-api-access-gvf7v" (OuterVolumeSpecName: "kube-api-access-gvf7v") pod "3854c241-241b-49c9-a095-bac632e89d5b" (UID: "3854c241-241b-49c9-a095-bac632e89d5b"). InnerVolumeSpecName "kube-api-access-gvf7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.662432 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-scripts" (OuterVolumeSpecName: "scripts") pod "3854c241-241b-49c9-a095-bac632e89d5b" (UID: "3854c241-241b-49c9-a095-bac632e89d5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.738899 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.738929 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3854c241-241b-49c9-a095-bac632e89d5b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.738938 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.738948 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvf7v\" (UniqueName: \"kubernetes.io/projected/3854c241-241b-49c9-a095-bac632e89d5b-kube-api-access-gvf7v\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.776957 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.777238 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3854c241-241b-49c9-a095-bac632e89d5b","Type":"ContainerDied","Data":"3330d196c0ffa687cce8d8c66c958efeddedbdc4d75d93d46b3393cd0f75c6eb"} Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.777303 4678 scope.go:117] "RemoveContainer" containerID="512972e6f15d7024be99291c48b0a98b7989b536416addb0c092de8879cb1098" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.804991 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c6f6bc877-zxzjz" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.805157 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-dbfccf5c9-tfx4w" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.805204 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-dbfccf5c9-tfx4w" event={"ID":"6faea971-f662-49d1-b48c-5743f127d7b5","Type":"ContainerDied","Data":"ed587d7c0cc1a4e3a0047c2e65b4e67cb65ed7cfc5a2576b4c90599f888ddca7"} Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.805350 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8b488c65-nmwlp" Dec 06 10:56:47 crc kubenswrapper[4678]: E1206 10:56:47.839417 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc939ac67_2678_47b0_9a57_7a10f89799bd.slice/crio-conmon-a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712.scope\": RecentStats: unable to find data in memory cache]" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.890987 4678 scope.go:117] "RemoveContainer" containerID="d7cb112822079b5ada3b60642d202ec2cb12e813e32e4c0b22bdc86a8f2c076c" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.902384 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c6f6bc877-zxzjz"] Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.920618 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c6f6bc877-zxzjz"] Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.947215 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-dbfccf5c9-tfx4w"] Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.950852 4678 scope.go:117] "RemoveContainer" containerID="fd37d0f403224fe0b9ae5fdd56a3b0dd389c3b987d4dfe06f0a769ded18b7265" Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.961614 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-dbfccf5c9-tfx4w"] Dec 06 10:56:47 crc kubenswrapper[4678]: I1206 10:56:47.977844 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b8b488c65-nmwlp"] Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.007012 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b8b488c65-nmwlp"] Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.029948 4678 scope.go:117] "RemoveContainer" containerID="3623d2ee1b98fd054f950002a6caaa770eb1c83a5b39063c09a7899ec4d7a50c" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.120616 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3854c241-241b-49c9-a095-bac632e89d5b" (UID: "3854c241-241b-49c9-a095-bac632e89d5b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.150965 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.380577 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3854c241-241b-49c9-a095-bac632e89d5b" (UID: "3854c241-241b-49c9-a095-bac632e89d5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.414945 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-config-data" (OuterVolumeSpecName: "config-data") pod "3854c241-241b-49c9-a095-bac632e89d5b" (UID: "3854c241-241b-49c9-a095-bac632e89d5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.481675 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.481709 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3854c241-241b-49c9-a095-bac632e89d5b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.484925 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.631962 4678 scope.go:117] "RemoveContainer" containerID="1cc46eeca1eaffbcb92eb5d37e993d0510583804e6a803421637bf4b0a388830" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671289 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7b874f6c6d-jcklm"] Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671747 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="proxy-httpd" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671769 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="proxy-httpd" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671783 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671789 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671799 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671805 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671811 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671818 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671828 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="ceilometer-notification-agent" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671833 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="ceilometer-notification-agent" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671847 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671853 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671868 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de91d1f3-4bc4-4611-ad00-169a5711fd54" containerName="init" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671873 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="de91d1f3-4bc4-4611-ad00-169a5711fd54" containerName="init" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671887 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="sg-core" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671892 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="sg-core" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671903 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-httpd" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671908 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-httpd" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671922 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671928 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671935 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671940 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.671955 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-api" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.671961 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-api" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672155 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="ceilometer-notification-agent" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672174 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672181 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672190 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="proxy-httpd" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672199 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672208 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-httpd" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672213 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672222 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672230 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="de91d1f3-4bc4-4611-ad00-169a5711fd54" containerName="init" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672240 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3854c241-241b-49c9-a095-bac632e89d5b" containerName="sg-core" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672248 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" containerName="horizon-log" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.672259 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerName="neutron-api" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.673266 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.684470 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-ovndb-tls-certs\") pod \"c939ac67-2678-47b0-9a57-7a10f89799bd\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.684673 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-config\") pod \"c939ac67-2678-47b0-9a57-7a10f89799bd\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.684734 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-combined-ca-bundle\") pod \"c939ac67-2678-47b0-9a57-7a10f89799bd\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.684754 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-httpd-config\") pod \"c939ac67-2678-47b0-9a57-7a10f89799bd\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.684805 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2ccb\" (UniqueName: \"kubernetes.io/projected/c939ac67-2678-47b0-9a57-7a10f89799bd-kube-api-access-h2ccb\") pod \"c939ac67-2678-47b0-9a57-7a10f89799bd\" (UID: \"c939ac67-2678-47b0-9a57-7a10f89799bd\") " Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.686053 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.686257 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.689058 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b874f6c6d-jcklm"] Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.717340 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c939ac67-2678-47b0-9a57-7a10f89799bd-kube-api-access-h2ccb" (OuterVolumeSpecName: "kube-api-access-h2ccb") pod "c939ac67-2678-47b0-9a57-7a10f89799bd" (UID: "c939ac67-2678-47b0-9a57-7a10f89799bd"). InnerVolumeSpecName "kube-api-access-h2ccb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.778997 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "c939ac67-2678-47b0-9a57-7a10f89799bd" (UID: "c939ac67-2678-47b0-9a57-7a10f89799bd"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791598 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfw6w\" (UniqueName: \"kubernetes.io/projected/6e0b0cc3-b79e-4e3a-8459-adba09620269-kube-api-access-lfw6w\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791640 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-internal-tls-certs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791658 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0b0cc3-b79e-4e3a-8459-adba09620269-logs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791706 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-config-data-custom\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791720 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-combined-ca-bundle\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791736 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-public-tls-certs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791797 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-config-data\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791856 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.791868 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2ccb\" (UniqueName: \"kubernetes.io/projected/c939ac67-2678-47b0-9a57-7a10f89799bd-kube-api-access-h2ccb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.803677 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-config" (OuterVolumeSpecName: "config") pod "c939ac67-2678-47b0-9a57-7a10f89799bd" (UID: "c939ac67-2678-47b0-9a57-7a10f89799bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.837451 4678 generic.go:334] "Generic (PLEG): container finished" podID="c939ac67-2678-47b0-9a57-7a10f89799bd" containerID="a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712" exitCode=0 Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.837767 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4858bb8d-jvvsq" event={"ID":"c939ac67-2678-47b0-9a57-7a10f89799bd","Type":"ContainerDied","Data":"a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.838299 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f4858bb8d-jvvsq" event={"ID":"c939ac67-2678-47b0-9a57-7a10f89799bd","Type":"ContainerDied","Data":"1133de2a92c0dbc1b7e9131bea8061275eb1d3fd468d0939e8e44e3dbfa668c5"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.839363 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5674df646-s4z6g" event={"ID":"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b","Type":"ContainerStarted","Data":"099fae3f184dfd2b5aa1867c9d1d70291b0f31bce68214deb248ef5885d23de6"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.838628 4678 scope.go:117] "RemoveContainer" containerID="7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.838189 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f4858bb8d-jvvsq" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.852583 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-597bfcdc9c-vdzb7" event={"ID":"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957","Type":"ContainerStarted","Data":"0b2aa07759687e3c94c7fd397c026b6c676eaf86d423672989a71ec341027256"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.853946 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-597bfcdc9c-vdzb7" event={"ID":"69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957","Type":"ContainerStarted","Data":"893cf4f09f1c1c14565ec227a9f52e4bfb4bfe3c92b089ff25744b56922fea44"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.867399 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c939ac67-2678-47b0-9a57-7a10f89799bd" (UID: "c939ac67-2678-47b0-9a57-7a10f89799bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.882439 4678 scope.go:117] "RemoveContainer" containerID="a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.888864 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d4ece9-a8c0-4842-a3d3-d5c05c77676b","Type":"ContainerStarted","Data":"d0eb0137b8e92d5cf5437f09289085a181f428fa4b015276e1e04690908dc91c"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.889007 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api-log" containerID="cri-o://1cc7475c58e76a748f418936dc0f0af1b64deac6832d1a31e3f736f45e4011cc" gracePeriod=30 Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.889238 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.889271 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api" containerID="cri-o://d0eb0137b8e92d5cf5437f09289085a181f428fa4b015276e1e04690908dc91c" gracePeriod=30 Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893231 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-config-data\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893300 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfw6w\" (UniqueName: \"kubernetes.io/projected/6e0b0cc3-b79e-4e3a-8459-adba09620269-kube-api-access-lfw6w\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893326 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-internal-tls-certs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893340 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0b0cc3-b79e-4e3a-8459-adba09620269-logs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893385 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-config-data-custom\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893400 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-combined-ca-bundle\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893418 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-public-tls-certs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893460 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.893472 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.901415 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "c939ac67-2678-47b0-9a57-7a10f89799bd" (UID: "c939ac67-2678-47b0-9a57-7a10f89799bd"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.906084 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e0b0cc3-b79e-4e3a-8459-adba09620269-logs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.914559 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-public-tls-certs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.915356 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-combined-ca-bundle\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.915390 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.917341 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-config-data-custom\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.917449 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-internal-tls-certs\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.920670 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e0b0cc3-b79e-4e3a-8459-adba09620269-config-data\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.926779 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" event={"ID":"eb1e07f0-3034-47d2-9143-82058df6d73c","Type":"ContainerStarted","Data":"d554c6311667f0da84e4fa15eb86691184370fbf4716480b062539f4ef603e82"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.927532 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.946904 4678 scope.go:117] "RemoveContainer" containerID="7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.959392 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae\": container with ID starting with 7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae not found: ID does not exist" containerID="7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.959725 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae"} err="failed to get container status \"7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae\": rpc error: code = NotFound desc = could not find container \"7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae\": container with ID starting with 7a54cd619a47fa5a7595d932615575a240439787ee49aa62123dc030f6548cae not found: ID does not exist" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.959753 4678 scope.go:117] "RemoveContainer" containerID="a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712" Dec 06 10:56:48 crc kubenswrapper[4678]: E1206 10:56:48.970907 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712\": container with ID starting with a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712 not found: ID does not exist" containerID="a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.970942 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712"} err="failed to get container status \"a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712\": rpc error: code = NotFound desc = could not find container \"a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712\": container with ID starting with a463c6f7ea7e87a4dfcc96216c31ad609b6df020766a0d4c8a9a918f8ca1d712 not found: ID does not exist" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.975093 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.977105 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfw6w\" (UniqueName: \"kubernetes.io/projected/6e0b0cc3-b79e-4e3a-8459-adba09620269-kube-api-access-lfw6w\") pod \"barbican-api-7b874f6c6d-jcklm\" (UID: \"6e0b0cc3-b79e-4e3a-8459-adba09620269\") " pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.980772 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.980886 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.981608 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"389817da78af88665e5f2000c353529adec4c3df99e02f2713ce867a856272b2"} pod="openstack/horizon-5b97647d4d-lzbd6" containerMessage="Container horizon failed startup probe, will be restarted" Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.981694 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" containerID="cri-o://389817da78af88665e5f2000c353529adec4c3df99e02f2713ce867a856272b2" gracePeriod=30 Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.988691 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd48a1a0-2fe4-423e-8930-99d5347e218e","Type":"ContainerStarted","Data":"87c94b76d0950d17ac6152c20949bbdba9697d7200d998a48ac9636d18d98be7"} Dec 06 10:56:48 crc kubenswrapper[4678]: I1206 10:56:48.994842 4678 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c939ac67-2678-47b0-9a57-7a10f89799bd-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.005374 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.007873 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.021404 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-597bfcdc9c-vdzb7" podStartSLOduration=4.919072726 podStartE2EDuration="10.021382729s" podCreationTimestamp="2025-12-06 10:56:39 +0000 UTC" firstStartedPulling="2025-12-06 10:56:41.664995183 +0000 UTC m=+1206.508426622" lastFinishedPulling="2025-12-06 10:56:46.767305186 +0000 UTC m=+1211.610736625" observedRunningTime="2025-12-06 10:56:48.958913452 +0000 UTC m=+1213.802344881" watchObservedRunningTime="2025-12-06 10:56:49.021382729 +0000 UTC m=+1213.864814168" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.022375 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.022696 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.059555 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.060257 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" podStartSLOduration=9.060237334 podStartE2EDuration="9.060237334s" podCreationTimestamp="2025-12-06 10:56:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:49.023039969 +0000 UTC m=+1213.866471398" watchObservedRunningTime="2025-12-06 10:56:49.060237334 +0000 UTC m=+1213.903668773" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.078807 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.096014 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-log-httpd\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.096101 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-config-data\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.096125 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.096287 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-scripts\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.096322 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-run-httpd\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.096351 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52rk7\" (UniqueName: \"kubernetes.io/projected/a3899eb2-a276-4bc7-97d5-268ea406829a-kube-api-access-52rk7\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.096370 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.100742 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=8.100720974 podStartE2EDuration="8.100720974s" podCreationTimestamp="2025-12-06 10:56:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:49.080511759 +0000 UTC m=+1213.923943208" watchObservedRunningTime="2025-12-06 10:56:49.100720974 +0000 UTC m=+1213.944152403" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.197560 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-scripts\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.197618 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-run-httpd\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.197644 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52rk7\" (UniqueName: \"kubernetes.io/projected/a3899eb2-a276-4bc7-97d5-268ea406829a-kube-api-access-52rk7\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.197665 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.197713 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-log-httpd\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.197740 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-config-data\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.197763 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.198719 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-log-httpd\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.205215 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-run-httpd\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.214090 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.216269 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.233371 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-scripts\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.236297 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-config-data\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.254989 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=8.514549575 podStartE2EDuration="9.254967594s" podCreationTimestamp="2025-12-06 10:56:40 +0000 UTC" firstStartedPulling="2025-12-06 10:56:42.07979937 +0000 UTC m=+1206.923230809" lastFinishedPulling="2025-12-06 10:56:42.820217389 +0000 UTC m=+1207.663648828" observedRunningTime="2025-12-06 10:56:49.23207257 +0000 UTC m=+1214.075504009" watchObservedRunningTime="2025-12-06 10:56:49.254967594 +0000 UTC m=+1214.098399033" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.255919 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52rk7\" (UniqueName: \"kubernetes.io/projected/a3899eb2-a276-4bc7-97d5-268ea406829a-kube-api-access-52rk7\") pod \"ceilometer-0\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.309192 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.321538 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5f4858bb8d-jvvsq"] Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.325885 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5f4858bb8d-jvvsq"] Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.496060 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c9e9db-8842-4412-9d09-4f8fdb8a39cf" path="/var/lib/kubelet/pods/13c9e9db-8842-4412-9d09-4f8fdb8a39cf/volumes" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.496955 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3854c241-241b-49c9-a095-bac632e89d5b" path="/var/lib/kubelet/pods/3854c241-241b-49c9-a095-bac632e89d5b/volumes" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.498414 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66418cf7-c293-4ddc-948c-60666d9fd1c4" path="/var/lib/kubelet/pods/66418cf7-c293-4ddc-948c-60666d9fd1c4/volumes" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.499174 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6faea971-f662-49d1-b48c-5743f127d7b5" path="/var/lib/kubelet/pods/6faea971-f662-49d1-b48c-5743f127d7b5/volumes" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.499813 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c939ac67-2678-47b0-9a57-7a10f89799bd" path="/var/lib/kubelet/pods/c939ac67-2678-47b0-9a57-7a10f89799bd/volumes" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.506783 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.881610 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:56:49 crc kubenswrapper[4678]: I1206 10:56:49.915174 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7b874f6c6d-jcklm"] Dec 06 10:56:50 crc kubenswrapper[4678]: I1206 10:56:50.002997 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerStarted","Data":"ddfe9dff086fd3f96f6c0a043693db91ca6d7cbc3f186a93529e5654c09775e7"} Dec 06 10:56:50 crc kubenswrapper[4678]: I1206 10:56:50.014854 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5674df646-s4z6g" event={"ID":"e55dbe47-fb3b-4020-bba4-f8c038e7cc5b","Type":"ContainerStarted","Data":"c9f0c6d601c544f6e7195f386723a81e071e587924d971f042a4bc50fc019b7c"} Dec 06 10:56:50 crc kubenswrapper[4678]: I1206 10:56:50.018840 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b874f6c6d-jcklm" event={"ID":"6e0b0cc3-b79e-4e3a-8459-adba09620269","Type":"ContainerStarted","Data":"ce8a721b913dda8ac07b47dfadfef50eba2c8845a1c89149f38e5b8d50a8eb87"} Dec 06 10:56:50 crc kubenswrapper[4678]: I1206 10:56:50.029830 4678 generic.go:334] "Generic (PLEG): container finished" podID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerID="1cc7475c58e76a748f418936dc0f0af1b64deac6832d1a31e3f736f45e4011cc" exitCode=143 Dec 06 10:56:50 crc kubenswrapper[4678]: I1206 10:56:50.031084 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d4ece9-a8c0-4842-a3d3-d5c05c77676b","Type":"ContainerDied","Data":"1cc7475c58e76a748f418936dc0f0af1b64deac6832d1a31e3f736f45e4011cc"} Dec 06 10:56:50 crc kubenswrapper[4678]: I1206 10:56:50.045030 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5674df646-s4z6g" podStartSLOduration=6.022030385 podStartE2EDuration="11.045012714s" podCreationTimestamp="2025-12-06 10:56:39 +0000 UTC" firstStartedPulling="2025-12-06 10:56:41.743205074 +0000 UTC m=+1206.586636513" lastFinishedPulling="2025-12-06 10:56:46.766187403 +0000 UTC m=+1211.609618842" observedRunningTime="2025-12-06 10:56:50.031476746 +0000 UTC m=+1214.874908185" watchObservedRunningTime="2025-12-06 10:56:50.045012714 +0000 UTC m=+1214.888444153" Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.040091 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerStarted","Data":"34fd7e8c1e90567ccb80db12934eff1b8e0fa9a5461d9d816216d1a8d3078d05"} Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.043832 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b874f6c6d-jcklm" event={"ID":"6e0b0cc3-b79e-4e3a-8459-adba09620269","Type":"ContainerStarted","Data":"8e63d67730e99aa99cc94ef0859d884d4e95e8b1c5c57581d44ed15c3d528ca6"} Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.043861 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7b874f6c6d-jcklm" event={"ID":"6e0b0cc3-b79e-4e3a-8459-adba09620269","Type":"ContainerStarted","Data":"652bc72af7fbf82af24b3303009a9367ea0d5db03f0575c6dbd18d4d76bb152e"} Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.044314 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.044342 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.047821 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.050199 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.159:8080/\": dial tcp 10.217.0.159:8080: connect: connection refused" Dec 06 10:56:51 crc kubenswrapper[4678]: I1206 10:56:51.079453 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7b874f6c6d-jcklm" podStartSLOduration=3.079435875 podStartE2EDuration="3.079435875s" podCreationTimestamp="2025-12-06 10:56:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:56:51.062741623 +0000 UTC m=+1215.906173062" watchObservedRunningTime="2025-12-06 10:56:51.079435875 +0000 UTC m=+1215.922867314" Dec 06 10:56:52 crc kubenswrapper[4678]: I1206 10:56:52.058548 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerStarted","Data":"8ec29306695acf64976687bd33d1b7b7cdb65d39f054dae502cc63f0df3acca4"} Dec 06 10:56:53 crc kubenswrapper[4678]: I1206 10:56:53.067823 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerStarted","Data":"8d72408cbe1c554b7a686099ac7000b22ad083e74ca16a08c681e503ce0271b0"} Dec 06 10:56:53 crc kubenswrapper[4678]: I1206 10:56:53.524914 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 06 10:56:54 crc kubenswrapper[4678]: I1206 10:56:54.016709 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:54 crc kubenswrapper[4678]: I1206 10:56:54.519409 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:56:55 crc kubenswrapper[4678]: I1206 10:56:55.084526 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerStarted","Data":"fa322abf4a1da61625f9e9e53f83b4d0485375c29fb3ef688ac6c33a91ca5a9f"} Dec 06 10:56:55 crc kubenswrapper[4678]: I1206 10:56:55.085559 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 10:56:55 crc kubenswrapper[4678]: I1206 10:56:55.119564 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.1414694499999998 podStartE2EDuration="7.119548658s" podCreationTimestamp="2025-12-06 10:56:48 +0000 UTC" firstStartedPulling="2025-12-06 10:56:49.895950317 +0000 UTC m=+1214.739381746" lastFinishedPulling="2025-12-06 10:56:53.874029515 +0000 UTC m=+1218.717460954" observedRunningTime="2025-12-06 10:56:55.10703449 +0000 UTC m=+1219.950465929" watchObservedRunningTime="2025-12-06 10:56:55.119548658 +0000 UTC m=+1219.962980097" Dec 06 10:56:56 crc kubenswrapper[4678]: I1206 10:56:56.332646 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:56:56 crc kubenswrapper[4678]: I1206 10:56:56.402916 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-2f2r2"] Dec 06 10:56:56 crc kubenswrapper[4678]: I1206 10:56:56.403304 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" podUID="0510beed-12f3-4505-9aec-cb6563d35885" containerName="dnsmasq-dns" containerID="cri-o://956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc" gracePeriod=10 Dec 06 10:56:56 crc kubenswrapper[4678]: I1206 10:56:56.941943 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 10:56:56 crc kubenswrapper[4678]: I1206 10:56:56.996187 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.118636 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.119807 4678 generic.go:334] "Generic (PLEG): container finished" podID="0510beed-12f3-4505-9aec-cb6563d35885" containerID="956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc" exitCode=0 Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.120298 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" event={"ID":"0510beed-12f3-4505-9aec-cb6563d35885","Type":"ContainerDied","Data":"956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc"} Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.120345 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" event={"ID":"0510beed-12f3-4505-9aec-cb6563d35885","Type":"ContainerDied","Data":"1021d77d737c50ee54b270dbd1e418ec9b992485f039039bbb753e3e97d8e100"} Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.120368 4678 scope.go:117] "RemoveContainer" containerID="956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.120630 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="cinder-scheduler" containerID="cri-o://7c2b7044c50faf48f6df34973d69856dbf7569ad12ced58571ac9ff8f41badb6" gracePeriod=30 Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.120720 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="probe" containerID="cri-o://87c94b76d0950d17ac6152c20949bbdba9697d7200d998a48ac9636d18d98be7" gracePeriod=30 Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.180406 4678 scope.go:117] "RemoveContainer" containerID="b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.188020 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-svc\") pod \"0510beed-12f3-4505-9aec-cb6563d35885\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.188064 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzz59\" (UniqueName: \"kubernetes.io/projected/0510beed-12f3-4505-9aec-cb6563d35885-kube-api-access-nzz59\") pod \"0510beed-12f3-4505-9aec-cb6563d35885\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.188131 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-nb\") pod \"0510beed-12f3-4505-9aec-cb6563d35885\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.188175 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-swift-storage-0\") pod \"0510beed-12f3-4505-9aec-cb6563d35885\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.188222 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-sb\") pod \"0510beed-12f3-4505-9aec-cb6563d35885\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.188370 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-config\") pod \"0510beed-12f3-4505-9aec-cb6563d35885\" (UID: \"0510beed-12f3-4505-9aec-cb6563d35885\") " Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.224686 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0510beed-12f3-4505-9aec-cb6563d35885-kube-api-access-nzz59" (OuterVolumeSpecName: "kube-api-access-nzz59") pod "0510beed-12f3-4505-9aec-cb6563d35885" (UID: "0510beed-12f3-4505-9aec-cb6563d35885"). InnerVolumeSpecName "kube-api-access-nzz59". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.297629 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzz59\" (UniqueName: \"kubernetes.io/projected/0510beed-12f3-4505-9aec-cb6563d35885-kube-api-access-nzz59\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.341277 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0510beed-12f3-4505-9aec-cb6563d35885" (UID: "0510beed-12f3-4505-9aec-cb6563d35885"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.354446 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0510beed-12f3-4505-9aec-cb6563d35885" (UID: "0510beed-12f3-4505-9aec-cb6563d35885"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.355106 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0510beed-12f3-4505-9aec-cb6563d35885" (UID: "0510beed-12f3-4505-9aec-cb6563d35885"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.367755 4678 scope.go:117] "RemoveContainer" containerID="956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc" Dec 06 10:56:57 crc kubenswrapper[4678]: E1206 10:56:57.369065 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc\": container with ID starting with 956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc not found: ID does not exist" containerID="956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.369112 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc"} err="failed to get container status \"956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc\": rpc error: code = NotFound desc = could not find container \"956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc\": container with ID starting with 956f0bb52e9faead9f4fabd88a551fcce6cc0e0077067d1a745d767946fae3fc not found: ID does not exist" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.369141 4678 scope.go:117] "RemoveContainer" containerID="b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae" Dec 06 10:56:57 crc kubenswrapper[4678]: E1206 10:56:57.370068 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae\": container with ID starting with b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae not found: ID does not exist" containerID="b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.370102 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae"} err="failed to get container status \"b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae\": rpc error: code = NotFound desc = could not find container \"b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae\": container with ID starting with b5b7ece2100cd991b9026a93d702b174fe3c0d807b763e4aa7110f6e6b9522ae not found: ID does not exist" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.393978 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-config" (OuterVolumeSpecName: "config") pod "0510beed-12f3-4505-9aec-cb6563d35885" (UID: "0510beed-12f3-4505-9aec-cb6563d35885"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.400590 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.400619 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.400628 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.400639 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.406645 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0510beed-12f3-4505-9aec-cb6563d35885" (UID: "0510beed-12f3-4505-9aec-cb6563d35885"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.501941 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0510beed-12f3-4505-9aec-cb6563d35885-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.617252 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:57 crc kubenswrapper[4678]: I1206 10:56:57.926160 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5bd747c594-kkr56" Dec 06 10:56:58 crc kubenswrapper[4678]: I1206 10:56:58.148929 4678 generic.go:334] "Generic (PLEG): container finished" podID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerID="87c94b76d0950d17ac6152c20949bbdba9697d7200d998a48ac9636d18d98be7" exitCode=0 Dec 06 10:56:58 crc kubenswrapper[4678]: I1206 10:56:58.148982 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd48a1a0-2fe4-423e-8930-99d5347e218e","Type":"ContainerDied","Data":"87c94b76d0950d17ac6152c20949bbdba9697d7200d998a48ac9636d18d98be7"} Dec 06 10:56:58 crc kubenswrapper[4678]: I1206 10:56:58.155233 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-2f2r2" Dec 06 10:56:58 crc kubenswrapper[4678]: I1206 10:56:58.201273 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-2f2r2"] Dec 06 10:56:58 crc kubenswrapper[4678]: I1206 10:56:58.232338 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-2f2r2"] Dec 06 10:56:58 crc kubenswrapper[4678]: E1206 10:56:58.248174 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd48a1a0_2fe4_423e_8930_99d5347e218e.slice/crio-87c94b76d0950d17ac6152c20949bbdba9697d7200d998a48ac9636d18d98be7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd48a1a0_2fe4_423e_8930_99d5347e218e.slice/crio-conmon-87c94b76d0950d17ac6152c20949bbdba9697d7200d998a48ac9636d18d98be7.scope\": RecentStats: unable to find data in memory cache]" Dec 06 10:56:58 crc kubenswrapper[4678]: I1206 10:56:58.519062 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7f69d54bd9-rd99t" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.315039 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.315408 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.316754 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"a04f288a96a3e3ff2987759b3e148e1cbbffe2e84ab32bb6357b33ca748488f8"} pod="openstack/horizon-7d77bcbbf8-cfc4l" containerMessage="Container horizon failed startup probe, will be restarted" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.316807 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" containerName="horizon" containerID="cri-o://a04f288a96a3e3ff2987759b3e148e1cbbffe2e84ab32bb6357b33ca748488f8" gracePeriod=30 Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.335830 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 06 10:56:59 crc kubenswrapper[4678]: E1206 10:56:59.336567 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0510beed-12f3-4505-9aec-cb6563d35885" containerName="init" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.336584 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0510beed-12f3-4505-9aec-cb6563d35885" containerName="init" Dec 06 10:56:59 crc kubenswrapper[4678]: E1206 10:56:59.336635 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0510beed-12f3-4505-9aec-cb6563d35885" containerName="dnsmasq-dns" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.336642 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0510beed-12f3-4505-9aec-cb6563d35885" containerName="dnsmasq-dns" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.336981 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="0510beed-12f3-4505-9aec-cb6563d35885" containerName="dnsmasq-dns" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.337895 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.349443 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.349743 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.349960 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zs8sj" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.355368 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.467721 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66756fbd-c49f-4ea6-8595-7801a7ddecf3-openstack-config-secret\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.467786 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66756fbd-c49f-4ea6-8595-7801a7ddecf3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.467854 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66756fbd-c49f-4ea6-8595-7801a7ddecf3-openstack-config\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.467894 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l86kz\" (UniqueName: \"kubernetes.io/projected/66756fbd-c49f-4ea6-8595-7801a7ddecf3-kube-api-access-l86kz\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.503727 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0510beed-12f3-4505-9aec-cb6563d35885" path="/var/lib/kubelet/pods/0510beed-12f3-4505-9aec-cb6563d35885/volumes" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.560719 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.569512 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66756fbd-c49f-4ea6-8595-7801a7ddecf3-openstack-config-secret\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.569864 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66756fbd-c49f-4ea6-8595-7801a7ddecf3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.569977 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66756fbd-c49f-4ea6-8595-7801a7ddecf3-openstack-config\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.570073 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l86kz\" (UniqueName: \"kubernetes.io/projected/66756fbd-c49f-4ea6-8595-7801a7ddecf3-kube-api-access-l86kz\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.570948 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/66756fbd-c49f-4ea6-8595-7801a7ddecf3-openstack-config\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.584698 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66756fbd-c49f-4ea6-8595-7801a7ddecf3-combined-ca-bundle\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.585020 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/66756fbd-c49f-4ea6-8595-7801a7ddecf3-openstack-config-secret\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.588988 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l86kz\" (UniqueName: \"kubernetes.io/projected/66756fbd-c49f-4ea6-8595-7801a7ddecf3-kube-api-access-l86kz\") pod \"openstackclient\" (UID: \"66756fbd-c49f-4ea6-8595-7801a7ddecf3\") " pod="openstack/openstackclient" Dec 06 10:56:59 crc kubenswrapper[4678]: I1206 10:56:59.687205 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 06 10:57:00 crc kubenswrapper[4678]: I1206 10:57:00.090632 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7b874f6c6d-jcklm" podUID="6e0b0cc3-b79e-4e3a-8459-adba09620269" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:57:00 crc kubenswrapper[4678]: I1206 10:57:00.090647 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7b874f6c6d-jcklm" podUID="6e0b0cc3-b79e-4e3a-8459-adba09620269" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:57:00 crc kubenswrapper[4678]: I1206 10:57:00.177834 4678 generic.go:334] "Generic (PLEG): container finished" podID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerID="7c2b7044c50faf48f6df34973d69856dbf7569ad12ced58571ac9ff8f41badb6" exitCode=0 Dec 06 10:57:00 crc kubenswrapper[4678]: I1206 10:57:00.177872 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd48a1a0-2fe4-423e-8930-99d5347e218e","Type":"ContainerDied","Data":"7c2b7044c50faf48f6df34973d69856dbf7569ad12ced58571ac9ff8f41badb6"} Dec 06 10:57:00 crc kubenswrapper[4678]: I1206 10:57:00.280978 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 06 10:57:00 crc kubenswrapper[4678]: I1206 10:57:00.753670 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:57:00 crc kubenswrapper[4678]: I1206 10:57:00.937949 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.126752 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data-custom\") pod \"cd48a1a0-2fe4-423e-8930-99d5347e218e\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.127135 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd48a1a0-2fe4-423e-8930-99d5347e218e-etc-machine-id\") pod \"cd48a1a0-2fe4-423e-8930-99d5347e218e\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.127198 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-scripts\") pod \"cd48a1a0-2fe4-423e-8930-99d5347e218e\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.127222 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9p64\" (UniqueName: \"kubernetes.io/projected/cd48a1a0-2fe4-423e-8930-99d5347e218e-kube-api-access-c9p64\") pod \"cd48a1a0-2fe4-423e-8930-99d5347e218e\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.127250 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-combined-ca-bundle\") pod \"cd48a1a0-2fe4-423e-8930-99d5347e218e\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.127269 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data\") pod \"cd48a1a0-2fe4-423e-8930-99d5347e218e\" (UID: \"cd48a1a0-2fe4-423e-8930-99d5347e218e\") " Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.127970 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd48a1a0-2fe4-423e-8930-99d5347e218e-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cd48a1a0-2fe4-423e-8930-99d5347e218e" (UID: "cd48a1a0-2fe4-423e-8930-99d5347e218e"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.135262 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd48a1a0-2fe4-423e-8930-99d5347e218e-kube-api-access-c9p64" (OuterVolumeSpecName: "kube-api-access-c9p64") pod "cd48a1a0-2fe4-423e-8930-99d5347e218e" (UID: "cd48a1a0-2fe4-423e-8930-99d5347e218e"). InnerVolumeSpecName "kube-api-access-c9p64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.163729 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cd48a1a0-2fe4-423e-8930-99d5347e218e" (UID: "cd48a1a0-2fe4-423e-8930-99d5347e218e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.178639 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-scripts" (OuterVolumeSpecName: "scripts") pod "cd48a1a0-2fe4-423e-8930-99d5347e218e" (UID: "cd48a1a0-2fe4-423e-8930-99d5347e218e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.218718 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"66756fbd-c49f-4ea6-8595-7801a7ddecf3","Type":"ContainerStarted","Data":"4a57ff6c70b7a884a5950016195a97933323b96b0b7babc9108072bda6494b6d"} Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.223796 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cd48a1a0-2fe4-423e-8930-99d5347e218e","Type":"ContainerDied","Data":"49a437635f70a0c63809e6624a73dab596f471dc4071070910c9fa234afb35f2"} Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.223883 4678 scope.go:117] "RemoveContainer" containerID="87c94b76d0950d17ac6152c20949bbdba9697d7200d998a48ac9636d18d98be7" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.224040 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.230760 4678 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cd48a1a0-2fe4-423e-8930-99d5347e218e-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.230781 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.230791 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9p64\" (UniqueName: \"kubernetes.io/projected/cd48a1a0-2fe4-423e-8930-99d5347e218e-kube-api-access-c9p64\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.230821 4678 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.261853 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd48a1a0-2fe4-423e-8930-99d5347e218e" (UID: "cd48a1a0-2fe4-423e-8930-99d5347e218e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.268399 4678 scope.go:117] "RemoveContainer" containerID="7c2b7044c50faf48f6df34973d69856dbf7569ad12ced58571ac9ff8f41badb6" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.307803 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data" (OuterVolumeSpecName: "config-data") pod "cd48a1a0-2fe4-423e-8930-99d5347e218e" (UID: "cd48a1a0-2fe4-423e-8930-99d5347e218e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.332790 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.332831 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd48a1a0-2fe4-423e-8930-99d5347e218e-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.551551 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.561347 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.584880 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:57:01 crc kubenswrapper[4678]: E1206 10:57:01.585282 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="probe" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.585300 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="probe" Dec 06 10:57:01 crc kubenswrapper[4678]: E1206 10:57:01.585322 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="cinder-scheduler" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.585328 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="cinder-scheduler" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.587669 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="probe" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.587697 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" containerName="cinder-scheduler" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.588820 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.591186 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.667128 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.740622 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.740668 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-config-data\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.740685 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-scripts\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.740773 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcrv4\" (UniqueName: \"kubernetes.io/projected/918bbdd7-35bb-457d-a29f-5c6e4f04b062-kube-api-access-mcrv4\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.740816 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/918bbdd7-35bb-457d-a29f-5c6e4f04b062-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.740889 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.756731 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.161:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.842275 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.842597 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-config-data\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.842619 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-scripts\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.842645 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcrv4\" (UniqueName: \"kubernetes.io/projected/918bbdd7-35bb-457d-a29f-5c6e4f04b062-kube-api-access-mcrv4\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.842665 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/918bbdd7-35bb-457d-a29f-5c6e4f04b062-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.842730 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.844649 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/918bbdd7-35bb-457d-a29f-5c6e4f04b062-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.850826 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.853230 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-config-data\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.864688 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.875271 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/918bbdd7-35bb-457d-a29f-5c6e4f04b062-scripts\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.885218 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcrv4\" (UniqueName: \"kubernetes.io/projected/918bbdd7-35bb-457d-a29f-5c6e4f04b062-kube-api-access-mcrv4\") pod \"cinder-scheduler-0\" (UID: \"918bbdd7-35bb-457d-a29f-5c6e4f04b062\") " pod="openstack/cinder-scheduler-0" Dec 06 10:57:01 crc kubenswrapper[4678]: I1206 10:57:01.951282 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 06 10:57:02 crc kubenswrapper[4678]: I1206 10:57:02.564741 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 06 10:57:02 crc kubenswrapper[4678]: W1206 10:57:02.572372 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod918bbdd7_35bb_457d_a29f_5c6e4f04b062.slice/crio-de3d0f0b739a90448f87adbac834ea5aa904a3ec622f27e7fb0adcf50b317644 WatchSource:0}: Error finding container de3d0f0b739a90448f87adbac834ea5aa904a3ec622f27e7fb0adcf50b317644: Status 404 returned error can't find the container with id de3d0f0b739a90448f87adbac834ea5aa904a3ec622f27e7fb0adcf50b317644 Dec 06 10:57:03 crc kubenswrapper[4678]: I1206 10:57:03.013586 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:57:03 crc kubenswrapper[4678]: I1206 10:57:03.264315 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"918bbdd7-35bb-457d-a29f-5c6e4f04b062","Type":"ContainerStarted","Data":"de3d0f0b739a90448f87adbac834ea5aa904a3ec622f27e7fb0adcf50b317644"} Dec 06 10:57:03 crc kubenswrapper[4678]: I1206 10:57:03.503364 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd48a1a0-2fe4-423e-8930-99d5347e218e" path="/var/lib/kubelet/pods/cd48a1a0-2fe4-423e-8930-99d5347e218e/volumes" Dec 06 10:57:04 crc kubenswrapper[4678]: I1206 10:57:04.085767 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7b874f6c6d-jcklm" podUID="6e0b0cc3-b79e-4e3a-8459-adba09620269" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:57:04 crc kubenswrapper[4678]: I1206 10:57:04.282795 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"918bbdd7-35bb-457d-a29f-5c6e4f04b062","Type":"ContainerStarted","Data":"85c4f716398969a67887382364cff532119a0480c8272da112958b7b47d75458"} Dec 06 10:57:04 crc kubenswrapper[4678]: I1206 10:57:04.282837 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"918bbdd7-35bb-457d-a29f-5c6e4f04b062","Type":"ContainerStarted","Data":"6538f755e6e76c901f2b24513c8b3325b7648a84bfcc57e640fe42151a86f248"} Dec 06 10:57:04 crc kubenswrapper[4678]: I1206 10:57:04.308881 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.308856525 podStartE2EDuration="3.308856525s" podCreationTimestamp="2025-12-06 10:57:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:04.304758574 +0000 UTC m=+1229.148190023" watchObservedRunningTime="2025-12-06 10:57:04.308856525 +0000 UTC m=+1229.152287964" Dec 06 10:57:05 crc kubenswrapper[4678]: I1206 10:57:05.103686 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7b874f6c6d-jcklm" podUID="6e0b0cc3-b79e-4e3a-8459-adba09620269" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:57:05 crc kubenswrapper[4678]: I1206 10:57:05.103699 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7b874f6c6d-jcklm" podUID="6e0b0cc3-b79e-4e3a-8459-adba09620269" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.162:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:57:06 crc kubenswrapper[4678]: I1206 10:57:06.171020 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 10:57:06 crc kubenswrapper[4678]: I1206 10:57:06.770815 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7b874f6c6d-jcklm" Dec 06 10:57:06 crc kubenswrapper[4678]: I1206 10:57:06.833745 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d978978c4-qsthb"] Dec 06 10:57:06 crc kubenswrapper[4678]: I1206 10:57:06.833968 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" containerID="cri-o://b0facf2a6703c559ab4d6fd9a8fe36d4168c234d46da2104728642e1cc47fff1" gracePeriod=30 Dec 06 10:57:06 crc kubenswrapper[4678]: I1206 10:57:06.834335 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api" containerID="cri-o://0964ceddf233314aea34efa77bab3f1761d93ba02156121900794bb4df1ede45" gracePeriod=30 Dec 06 10:57:06 crc kubenswrapper[4678]: I1206 10:57:06.955803 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 06 10:57:07 crc kubenswrapper[4678]: I1206 10:57:07.326077 4678 generic.go:334] "Generic (PLEG): container finished" podID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerID="b0facf2a6703c559ab4d6fd9a8fe36d4168c234d46da2104728642e1cc47fff1" exitCode=143 Dec 06 10:57:07 crc kubenswrapper[4678]: I1206 10:57:07.326124 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d978978c4-qsthb" event={"ID":"cd23d8eb-f06c-48e4-9319-aa699a260ca0","Type":"ContainerDied","Data":"b0facf2a6703c559ab4d6fd9a8fe36d4168c234d46da2104728642e1cc47fff1"} Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.181882 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-f5f494679-7bm9s"] Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.184211 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.186852 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.187651 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.189181 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.209933 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-f5f494679-7bm9s"] Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.319386 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-internal-tls-certs\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.320159 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh4wc\" (UniqueName: \"kubernetes.io/projected/b31f2128-db6e-498b-bd38-d0b1c41a4603-kube-api-access-xh4wc\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.320337 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-public-tls-certs\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.320400 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-combined-ca-bundle\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.320455 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b31f2128-db6e-498b-bd38-d0b1c41a4603-log-httpd\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.320505 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b31f2128-db6e-498b-bd38-d0b1c41a4603-run-httpd\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.320536 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b31f2128-db6e-498b-bd38-d0b1c41a4603-etc-swift\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.320571 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-config-data\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422197 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-internal-tls-certs\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422257 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh4wc\" (UniqueName: \"kubernetes.io/projected/b31f2128-db6e-498b-bd38-d0b1c41a4603-kube-api-access-xh4wc\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422313 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-public-tls-certs\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422343 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-combined-ca-bundle\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422372 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b31f2128-db6e-498b-bd38-d0b1c41a4603-log-httpd\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422391 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b31f2128-db6e-498b-bd38-d0b1c41a4603-run-httpd\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422411 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b31f2128-db6e-498b-bd38-d0b1c41a4603-etc-swift\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.422431 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-config-data\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.423480 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b31f2128-db6e-498b-bd38-d0b1c41a4603-run-httpd\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.424029 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b31f2128-db6e-498b-bd38-d0b1c41a4603-log-httpd\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.431130 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-combined-ca-bundle\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.431169 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-public-tls-certs\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.431966 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-config-data\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.432804 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b31f2128-db6e-498b-bd38-d0b1c41a4603-etc-swift\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.443317 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh4wc\" (UniqueName: \"kubernetes.io/projected/b31f2128-db6e-498b-bd38-d0b1c41a4603-kube-api-access-xh4wc\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.451161 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b31f2128-db6e-498b-bd38-d0b1c41a4603-internal-tls-certs\") pod \"swift-proxy-f5f494679-7bm9s\" (UID: \"b31f2128-db6e-498b-bd38-d0b1c41a4603\") " pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:09 crc kubenswrapper[4678]: I1206 10:57:09.517520 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:10 crc kubenswrapper[4678]: I1206 10:57:10.422937 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-f5f494679-7bm9s"] Dec 06 10:57:10 crc kubenswrapper[4678]: I1206 10:57:10.712451 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": dial tcp 10.217.0.158:9311: connect: connection refused" Dec 06 10:57:10 crc kubenswrapper[4678]: I1206 10:57:10.712550 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": dial tcp 10.217.0.158:9311: connect: connection refused" Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.078222 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.078512 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-central-agent" containerID="cri-o://34fd7e8c1e90567ccb80db12934eff1b8e0fa9a5461d9d816216d1a8d3078d05" gracePeriod=30 Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.078732 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="sg-core" containerID="cri-o://8d72408cbe1c554b7a686099ac7000b22ad083e74ca16a08c681e503ce0271b0" gracePeriod=30 Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.078721 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-notification-agent" containerID="cri-o://8ec29306695acf64976687bd33d1b7b7cdb65d39f054dae502cc63f0df3acca4" gracePeriod=30 Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.078894 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="proxy-httpd" containerID="cri-o://fa322abf4a1da61625f9e9e53f83b4d0485375c29fb3ef688ac6c33a91ca5a9f" gracePeriod=30 Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.105794 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.377180 4678 generic.go:334] "Generic (PLEG): container finished" podID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerID="0964ceddf233314aea34efa77bab3f1761d93ba02156121900794bb4df1ede45" exitCode=0 Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.377249 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d978978c4-qsthb" event={"ID":"cd23d8eb-f06c-48e4-9319-aa699a260ca0","Type":"ContainerDied","Data":"0964ceddf233314aea34efa77bab3f1761d93ba02156121900794bb4df1ede45"} Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.383110 4678 generic.go:334] "Generic (PLEG): container finished" podID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerID="fa322abf4a1da61625f9e9e53f83b4d0485375c29fb3ef688ac6c33a91ca5a9f" exitCode=0 Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.383278 4678 generic.go:334] "Generic (PLEG): container finished" podID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerID="8d72408cbe1c554b7a686099ac7000b22ad083e74ca16a08c681e503ce0271b0" exitCode=2 Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.383279 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerDied","Data":"fa322abf4a1da61625f9e9e53f83b4d0485375c29fb3ef688ac6c33a91ca5a9f"} Dec 06 10:57:11 crc kubenswrapper[4678]: I1206 10:57:11.383482 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerDied","Data":"8d72408cbe1c554b7a686099ac7000b22ad083e74ca16a08c681e503ce0271b0"} Dec 06 10:57:12 crc kubenswrapper[4678]: I1206 10:57:12.360365 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 06 10:57:12 crc kubenswrapper[4678]: I1206 10:57:12.405854 4678 generic.go:334] "Generic (PLEG): container finished" podID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerID="8ec29306695acf64976687bd33d1b7b7cdb65d39f054dae502cc63f0df3acca4" exitCode=0 Dec 06 10:57:12 crc kubenswrapper[4678]: I1206 10:57:12.405883 4678 generic.go:334] "Generic (PLEG): container finished" podID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerID="34fd7e8c1e90567ccb80db12934eff1b8e0fa9a5461d9d816216d1a8d3078d05" exitCode=0 Dec 06 10:57:12 crc kubenswrapper[4678]: I1206 10:57:12.405902 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerDied","Data":"8ec29306695acf64976687bd33d1b7b7cdb65d39f054dae502cc63f0df3acca4"} Dec 06 10:57:12 crc kubenswrapper[4678]: I1206 10:57:12.405928 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerDied","Data":"34fd7e8c1e90567ccb80db12934eff1b8e0fa9a5461d9d816216d1a8d3078d05"} Dec 06 10:57:15 crc kubenswrapper[4678]: I1206 10:57:15.711987 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": dial tcp 10.217.0.158:9311: connect: connection refused" Dec 06 10:57:15 crc kubenswrapper[4678]: I1206 10:57:15.712024 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6d978978c4-qsthb" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.158:9311/healthcheck\": dial tcp 10.217.0.158:9311: connect: connection refused" Dec 06 10:57:16 crc kubenswrapper[4678]: I1206 10:57:16.183644 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:57:16 crc kubenswrapper[4678]: I1206 10:57:16.184310 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-log" containerID="cri-o://58709e930d1cfb066f66ad4f8a5d037d014cd421a4650834889f677ea07f4b15" gracePeriod=30 Dec 06 10:57:16 crc kubenswrapper[4678]: I1206 10:57:16.184531 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-httpd" containerID="cri-o://6200b7dd3e61dde1e396ff21451edde208c22b656f3b16792eddc1031dc13ecb" gracePeriod=30 Dec 06 10:57:16 crc kubenswrapper[4678]: I1206 10:57:16.445304 4678 generic.go:334] "Generic (PLEG): container finished" podID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerID="58709e930d1cfb066f66ad4f8a5d037d014cd421a4650834889f677ea07f4b15" exitCode=143 Dec 06 10:57:16 crc kubenswrapper[4678]: I1206 10:57:16.445344 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ccee4c53-a485-4390-b65f-c2f55f2881ad","Type":"ContainerDied","Data":"58709e930d1cfb066f66ad4f8a5d037d014cd421a4650834889f677ea07f4b15"} Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.025304 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-8kxvz"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.026465 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.038891 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8kxvz"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.070594 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmkfh\" (UniqueName: \"kubernetes.io/projected/081454ce-456d-4cd3-bec3-b7a06034ab39-kube-api-access-xmkfh\") pod \"nova-api-db-create-8kxvz\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.070728 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081454ce-456d-4cd3-bec3-b7a06034ab39-operator-scripts\") pod \"nova-api-db-create-8kxvz\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.162306 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-4zkvd"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.164464 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.183413 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmkfh\" (UniqueName: \"kubernetes.io/projected/081454ce-456d-4cd3-bec3-b7a06034ab39-kube-api-access-xmkfh\") pod \"nova-api-db-create-8kxvz\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.183748 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081454ce-456d-4cd3-bec3-b7a06034ab39-operator-scripts\") pod \"nova-api-db-create-8kxvz\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.184505 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081454ce-456d-4cd3-bec3-b7a06034ab39-operator-scripts\") pod \"nova-api-db-create-8kxvz\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.222572 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-455a-account-create-update-kg2j5"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.224778 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.228019 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.287344 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stgp2\" (UniqueName: \"kubernetes.io/projected/d549492c-fcf7-402b-95e5-c0bee52f86eb-kube-api-access-stgp2\") pod \"nova-cell0-db-create-4zkvd\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.287722 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-operator-scripts\") pod \"nova-api-455a-account-create-update-kg2j5\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.287796 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55wsd\" (UniqueName: \"kubernetes.io/projected/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-kube-api-access-55wsd\") pod \"nova-api-455a-account-create-update-kg2j5\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.287819 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d549492c-fcf7-402b-95e5-c0bee52f86eb-operator-scripts\") pod \"nova-cell0-db-create-4zkvd\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.305968 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmkfh\" (UniqueName: \"kubernetes.io/projected/081454ce-456d-4cd3-bec3-b7a06034ab39-kube-api-access-xmkfh\") pod \"nova-api-db-create-8kxvz\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.326453 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4zkvd"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.372039 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.375627 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-455a-account-create-update-kg2j5"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.393082 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stgp2\" (UniqueName: \"kubernetes.io/projected/d549492c-fcf7-402b-95e5-c0bee52f86eb-kube-api-access-stgp2\") pod \"nova-cell0-db-create-4zkvd\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.393150 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-operator-scripts\") pod \"nova-api-455a-account-create-update-kg2j5\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.393215 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55wsd\" (UniqueName: \"kubernetes.io/projected/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-kube-api-access-55wsd\") pod \"nova-api-455a-account-create-update-kg2j5\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.393232 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d549492c-fcf7-402b-95e5-c0bee52f86eb-operator-scripts\") pod \"nova-cell0-db-create-4zkvd\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.394037 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d549492c-fcf7-402b-95e5-c0bee52f86eb-operator-scripts\") pod \"nova-cell0-db-create-4zkvd\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.394092 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-operator-scripts\") pod \"nova-api-455a-account-create-update-kg2j5\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.422134 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stgp2\" (UniqueName: \"kubernetes.io/projected/d549492c-fcf7-402b-95e5-c0bee52f86eb-kube-api-access-stgp2\") pod \"nova-cell0-db-create-4zkvd\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.437553 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55wsd\" (UniqueName: \"kubernetes.io/projected/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-kube-api-access-55wsd\") pod \"nova-api-455a-account-create-update-kg2j5\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.507566 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-6btts"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.508734 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.510944 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.569560 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6btts"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.607141 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-39d4-account-create-update-wpqkc"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.608400 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.608684 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a54c6d51-0b8e-491d-b5fb-8d08a260232a-operator-scripts\") pod \"nova-cell1-db-create-6btts\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.608941 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvtmr\" (UniqueName: \"kubernetes.io/projected/a54c6d51-0b8e-491d-b5fb-8d08a260232a-kube-api-access-pvtmr\") pod \"nova-cell1-db-create-6btts\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.612097 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.628224 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.640705 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-39d4-account-create-update-wpqkc"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.651865 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7391-account-create-update-74ks6"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.653079 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.665713 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.699158 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7391-account-create-update-74ks6"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.711640 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bd7n\" (UniqueName: \"kubernetes.io/projected/582eb659-cea1-44f9-9d87-a2d965a9bcfc-kube-api-access-5bd7n\") pod \"nova-cell1-7391-account-create-update-74ks6\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.711745 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3dd118-3077-4c96-b13b-edb34a16d944-operator-scripts\") pod \"nova-cell0-39d4-account-create-update-wpqkc\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.711772 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk8hh\" (UniqueName: \"kubernetes.io/projected/1c3dd118-3077-4c96-b13b-edb34a16d944-kube-api-access-tk8hh\") pod \"nova-cell0-39d4-account-create-update-wpqkc\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.711848 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582eb659-cea1-44f9-9d87-a2d965a9bcfc-operator-scripts\") pod \"nova-cell1-7391-account-create-update-74ks6\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.711903 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvtmr\" (UniqueName: \"kubernetes.io/projected/a54c6d51-0b8e-491d-b5fb-8d08a260232a-kube-api-access-pvtmr\") pod \"nova-cell1-db-create-6btts\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.711981 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a54c6d51-0b8e-491d-b5fb-8d08a260232a-operator-scripts\") pod \"nova-cell1-db-create-6btts\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.712907 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a54c6d51-0b8e-491d-b5fb-8d08a260232a-operator-scripts\") pod \"nova-cell1-db-create-6btts\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.734401 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvtmr\" (UniqueName: \"kubernetes.io/projected/a54c6d51-0b8e-491d-b5fb-8d08a260232a-kube-api-access-pvtmr\") pod \"nova-cell1-db-create-6btts\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.813526 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582eb659-cea1-44f9-9d87-a2d965a9bcfc-operator-scripts\") pod \"nova-cell1-7391-account-create-update-74ks6\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.813662 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bd7n\" (UniqueName: \"kubernetes.io/projected/582eb659-cea1-44f9-9d87-a2d965a9bcfc-kube-api-access-5bd7n\") pod \"nova-cell1-7391-account-create-update-74ks6\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.813702 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3dd118-3077-4c96-b13b-edb34a16d944-operator-scripts\") pod \"nova-cell0-39d4-account-create-update-wpqkc\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.813717 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk8hh\" (UniqueName: \"kubernetes.io/projected/1c3dd118-3077-4c96-b13b-edb34a16d944-kube-api-access-tk8hh\") pod \"nova-cell0-39d4-account-create-update-wpqkc\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.814203 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582eb659-cea1-44f9-9d87-a2d965a9bcfc-operator-scripts\") pod \"nova-cell1-7391-account-create-update-74ks6\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.814596 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3dd118-3077-4c96-b13b-edb34a16d944-operator-scripts\") pod \"nova-cell0-39d4-account-create-update-wpqkc\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.830683 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bd7n\" (UniqueName: \"kubernetes.io/projected/582eb659-cea1-44f9-9d87-a2d965a9bcfc-kube-api-access-5bd7n\") pod \"nova-cell1-7391-account-create-update-74ks6\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.838078 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk8hh\" (UniqueName: \"kubernetes.io/projected/1c3dd118-3077-4c96-b13b-edb34a16d944-kube-api-access-tk8hh\") pod \"nova-cell0-39d4-account-create-update-wpqkc\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.870380 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.946939 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.980735 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.981014 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-log" containerID="cri-o://abe07ac3ca1ab61beacbed1af31ef4a382dcd3d073c4b53028ae90606ff58fc7" gracePeriod=30 Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.981178 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-httpd" containerID="cri-o://7c08a0e53ba728e5e8bef5ac77f4984309f75903c6d92ea13330a8497fd15033" gracePeriod=30 Dec 06 10:57:17 crc kubenswrapper[4678]: I1206 10:57:17.993220 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:18 crc kubenswrapper[4678]: I1206 10:57:18.497189 4678 generic.go:334] "Generic (PLEG): container finished" podID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerID="abe07ac3ca1ab61beacbed1af31ef4a382dcd3d073c4b53028ae90606ff58fc7" exitCode=143 Dec 06 10:57:18 crc kubenswrapper[4678]: I1206 10:57:18.497314 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9aac132-0acd-48d8-a2e7-2d3b82157677","Type":"ContainerDied","Data":"abe07ac3ca1ab61beacbed1af31ef4a382dcd3d073c4b53028ae90606ff58fc7"} Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.518871 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.595770 4678 generic.go:334] "Generic (PLEG): container finished" podID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerID="d0eb0137b8e92d5cf5437f09289085a181f428fa4b015276e1e04690908dc91c" exitCode=137 Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.595827 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d4ece9-a8c0-4842-a3d3-d5c05c77676b","Type":"ContainerDied","Data":"d0eb0137b8e92d5cf5437f09289085a181f428fa4b015276e1e04690908dc91c"} Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.597296 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6d978978c4-qsthb" event={"ID":"cd23d8eb-f06c-48e4-9319-aa699a260ca0","Type":"ContainerDied","Data":"f0ddfe9b686bcb88e8e88100258841c73ac8aa405c4a7cb0162eafa7ab6a5c5f"} Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.597315 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0ddfe9b686bcb88e8e88100258841c73ac8aa405c4a7cb0162eafa7ab6a5c5f" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.598007 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f5f494679-7bm9s" event={"ID":"b31f2128-db6e-498b-bd38-d0b1c41a4603","Type":"ContainerStarted","Data":"054b084a80a51c59b7581a90930ee9b253a55a8918dde8b64538e5037d8948de"} Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.638957 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.639840 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a3899eb2-a276-4bc7-97d5-268ea406829a","Type":"ContainerDied","Data":"ddfe9dff086fd3f96f6c0a043693db91ca6d7cbc3f186a93529e5654c09775e7"} Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.639906 4678 scope.go:117] "RemoveContainer" containerID="fa322abf4a1da61625f9e9e53f83b4d0485375c29fb3ef688ac6c33a91ca5a9f" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.643041 4678 generic.go:334] "Generic (PLEG): container finished" podID="c3692143-6629-40ba-984b-187126e4a2ed" containerID="389817da78af88665e5f2000c353529adec4c3df99e02f2713ce867a856272b2" exitCode=137 Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.643088 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerDied","Data":"389817da78af88665e5f2000c353529adec4c3df99e02f2713ce867a856272b2"} Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.652406 4678 generic.go:334] "Generic (PLEG): container finished" podID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerID="6200b7dd3e61dde1e396ff21451edde208c22b656f3b16792eddc1031dc13ecb" exitCode=0 Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.652466 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ccee4c53-a485-4390-b65f-c2f55f2881ad","Type":"ContainerDied","Data":"6200b7dd3e61dde1e396ff21451edde208c22b656f3b16792eddc1031dc13ecb"} Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.664053 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.671398 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-combined-ca-bundle\") pod \"a3899eb2-a276-4bc7-97d5-268ea406829a\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.671462 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-sg-core-conf-yaml\") pod \"a3899eb2-a276-4bc7-97d5-268ea406829a\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.671646 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-scripts\") pod \"a3899eb2-a276-4bc7-97d5-268ea406829a\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.671693 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-run-httpd\") pod \"a3899eb2-a276-4bc7-97d5-268ea406829a\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.671712 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-config-data\") pod \"a3899eb2-a276-4bc7-97d5-268ea406829a\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.671816 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-log-httpd\") pod \"a3899eb2-a276-4bc7-97d5-268ea406829a\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.671833 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52rk7\" (UniqueName: \"kubernetes.io/projected/a3899eb2-a276-4bc7-97d5-268ea406829a-kube-api-access-52rk7\") pod \"a3899eb2-a276-4bc7-97d5-268ea406829a\" (UID: \"a3899eb2-a276-4bc7-97d5-268ea406829a\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.677657 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a3899eb2-a276-4bc7-97d5-268ea406829a" (UID: "a3899eb2-a276-4bc7-97d5-268ea406829a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.683728 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a3899eb2-a276-4bc7-97d5-268ea406829a" (UID: "a3899eb2-a276-4bc7-97d5-268ea406829a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.685910 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3899eb2-a276-4bc7-97d5-268ea406829a-kube-api-access-52rk7" (OuterVolumeSpecName: "kube-api-access-52rk7") pod "a3899eb2-a276-4bc7-97d5-268ea406829a" (UID: "a3899eb2-a276-4bc7-97d5-268ea406829a"). InnerVolumeSpecName "kube-api-access-52rk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.687840 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-8kxvz"] Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.714159 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6btts"] Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.725701 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-scripts" (OuterVolumeSpecName: "scripts") pod "a3899eb2-a276-4bc7-97d5-268ea406829a" (UID: "a3899eb2-a276-4bc7-97d5-268ea406829a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.772979 4678 scope.go:117] "RemoveContainer" containerID="8d72408cbe1c554b7a686099ac7000b22ad083e74ca16a08c681e503ce0271b0" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.773648 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-combined-ca-bundle\") pod \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.773795 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd23d8eb-f06c-48e4-9319-aa699a260ca0-logs\") pod \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.773822 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgvfz\" (UniqueName: \"kubernetes.io/projected/cd23d8eb-f06c-48e4-9319-aa699a260ca0-kube-api-access-jgvfz\") pod \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.773881 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data-custom\") pod \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.773964 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data\") pod \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\" (UID: \"cd23d8eb-f06c-48e4-9319-aa699a260ca0\") " Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.774318 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.774328 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.774337 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a3899eb2-a276-4bc7-97d5-268ea406829a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.774348 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52rk7\" (UniqueName: \"kubernetes.io/projected/a3899eb2-a276-4bc7-97d5-268ea406829a-kube-api-access-52rk7\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.774891 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd23d8eb-f06c-48e4-9319-aa699a260ca0-logs" (OuterVolumeSpecName: "logs") pod "cd23d8eb-f06c-48e4-9319-aa699a260ca0" (UID: "cd23d8eb-f06c-48e4-9319-aa699a260ca0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.800229 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "cd23d8eb-f06c-48e4-9319-aa699a260ca0" (UID: "cd23d8eb-f06c-48e4-9319-aa699a260ca0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.801350 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd23d8eb-f06c-48e4-9319-aa699a260ca0-kube-api-access-jgvfz" (OuterVolumeSpecName: "kube-api-access-jgvfz") pod "cd23d8eb-f06c-48e4-9319-aa699a260ca0" (UID: "cd23d8eb-f06c-48e4-9319-aa699a260ca0"). InnerVolumeSpecName "kube-api-access-jgvfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.877265 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd23d8eb-f06c-48e4-9319-aa699a260ca0-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.877299 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgvfz\" (UniqueName: \"kubernetes.io/projected/cd23d8eb-f06c-48e4-9319-aa699a260ca0-kube-api-access-jgvfz\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.877314 4678 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.896123 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-4zkvd"] Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.915929 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-39d4-account-create-update-wpqkc"] Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.946131 4678 scope.go:117] "RemoveContainer" containerID="8ec29306695acf64976687bd33d1b7b7cdb65d39f054dae502cc63f0df3acca4" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.978096 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd23d8eb-f06c-48e4-9319-aa699a260ca0" (UID: "cd23d8eb-f06c-48e4-9319-aa699a260ca0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:19 crc kubenswrapper[4678]: I1206 10:57:19.979397 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.002164 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a3899eb2-a276-4bc7-97d5-268ea406829a" (UID: "a3899eb2-a276-4bc7-97d5-268ea406829a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.059568 4678 scope.go:117] "RemoveContainer" containerID="34fd7e8c1e90567ccb80db12934eff1b8e0fa9a5461d9d816216d1a8d3078d05" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.082013 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.131876 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3899eb2-a276-4bc7-97d5-268ea406829a" (UID: "a3899eb2-a276-4bc7-97d5-268ea406829a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.189225 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data" (OuterVolumeSpecName: "config-data") pod "cd23d8eb-f06c-48e4-9319-aa699a260ca0" (UID: "cd23d8eb-f06c-48e4-9319-aa699a260ca0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.193667 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.194183 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd23d8eb-f06c-48e4-9319-aa699a260ca0-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.199721 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7391-account-create-update-74ks6"] Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.219817 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-455a-account-create-update-kg2j5"] Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.665606 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f5f494679-7bm9s" event={"ID":"b31f2128-db6e-498b-bd38-d0b1c41a4603","Type":"ContainerStarted","Data":"fcb8cfbb77ffa1191fc5b2349c3934e044dedec9b595e3f1300168c2ad930377"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.676538 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerStarted","Data":"c7685a17cd3751529dee7b1765aae00ffa7c22bf0283855f7f4a93abf135d645"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.682392 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"66756fbd-c49f-4ea6-8595-7801a7ddecf3","Type":"ContainerStarted","Data":"2de70e006ba1e03403836a6c42af989e7a1752d290fe1d53aa3457b6cfcffd95"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.684295 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" event={"ID":"1c3dd118-3077-4c96-b13b-edb34a16d944","Type":"ContainerStarted","Data":"f226fac5ff92d5521c8126125c7ea7b1f138f5f0939b61fbd9643a0c40b5741f"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.684323 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" event={"ID":"1c3dd118-3077-4c96-b13b-edb34a16d944","Type":"ContainerStarted","Data":"5ec3851341dcce486cfcfc56f923a3abef4f7681a830c5fd806c3698c9798cb6"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.687080 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6btts" event={"ID":"a54c6d51-0b8e-491d-b5fb-8d08a260232a","Type":"ContainerStarted","Data":"913a209132c2adb82e0be8f3c4fd4db2fefc66393db79043d32d41e49e286137"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.687106 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6btts" event={"ID":"a54c6d51-0b8e-491d-b5fb-8d08a260232a","Type":"ContainerStarted","Data":"f99befe25f574d09563702d36a8f61a3748ddac5be2961cafdc2933de2220144"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.689770 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-455a-account-create-update-kg2j5" event={"ID":"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee","Type":"ContainerStarted","Data":"be6ae9188bdb56aac9b830bcc6c1c51c5ab351ae2fee1236c3aaca1e6cf12c2c"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.694435 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ccee4c53-a485-4390-b65f-c2f55f2881ad","Type":"ContainerDied","Data":"77ef124e65a8722c021f47604fd0bf4b30df04bbe48dc2d14365b2c0bbaeb638"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.694476 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77ef124e65a8722c021f47604fd0bf4b30df04bbe48dc2d14365b2c0bbaeb638" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.701535 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4zkvd" event={"ID":"d549492c-fcf7-402b-95e5-c0bee52f86eb","Type":"ContainerStarted","Data":"6289bfa16d8e28be70e854c055f704746abd207cfd7a3d721df59face8a5b3c9"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.715975 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"23d4ece9-a8c0-4842-a3d3-d5c05c77676b","Type":"ContainerDied","Data":"d1422f9ae758ed0af0770176be16d4db093954bb2b3d91868e8274c427f39159"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.716016 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1422f9ae758ed0af0770176be16d4db093954bb2b3d91868e8274c427f39159" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.722988 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-config-data" (OuterVolumeSpecName: "config-data") pod "a3899eb2-a276-4bc7-97d5-268ea406829a" (UID: "a3899eb2-a276-4bc7-97d5-268ea406829a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.724547 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8kxvz" event={"ID":"081454ce-456d-4cd3-bec3-b7a06034ab39","Type":"ContainerStarted","Data":"1f8bf9573b476c1a5bea2e6927ac5c9c809a8683ed3bc8646ac581a40eca774a"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.728013 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6d978978c4-qsthb" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.728124 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7391-account-create-update-74ks6" event={"ID":"582eb659-cea1-44f9-9d87-a2d965a9bcfc","Type":"ContainerStarted","Data":"e02b7b915e3719a8f8f0f94bd71960314bb72332ad28120c0a31098157bb9bed"} Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.778682 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-6btts" podStartSLOduration=3.778663968 podStartE2EDuration="3.778663968s" podCreationTimestamp="2025-12-06 10:57:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:20.716120636 +0000 UTC m=+1245.559552065" watchObservedRunningTime="2025-12-06 10:57:20.778663968 +0000 UTC m=+1245.622095407" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.785338 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" podStartSLOduration=3.785318037 podStartE2EDuration="3.785318037s" podCreationTimestamp="2025-12-06 10:57:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:20.741055072 +0000 UTC m=+1245.584486511" watchObservedRunningTime="2025-12-06 10:57:20.785318037 +0000 UTC m=+1245.628749476" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.809530 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a3899eb2-a276-4bc7-97d5-268ea406829a-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.813186 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.229165244 podStartE2EDuration="21.813163505s" podCreationTimestamp="2025-12-06 10:56:59 +0000 UTC" firstStartedPulling="2025-12-06 10:57:00.344722157 +0000 UTC m=+1225.188153596" lastFinishedPulling="2025-12-06 10:57:18.928720418 +0000 UTC m=+1243.772151857" observedRunningTime="2025-12-06 10:57:20.759340079 +0000 UTC m=+1245.602771518" watchObservedRunningTime="2025-12-06 10:57:20.813163505 +0000 UTC m=+1245.656594944" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.848554 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.876901 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.894855 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6d978978c4-qsthb"] Dec 06 10:57:20 crc kubenswrapper[4678]: I1206 10:57:20.927566 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6d978978c4-qsthb"] Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.024475 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.024899 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data\") pod \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.024945 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-httpd-run\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.024999 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-scripts\") pod \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.025027 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-etc-machine-id\") pod \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.025071 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-public-tls-certs\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.025124 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-logs\") pod \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.025152 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdvbv\" (UniqueName: \"kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.025182 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-logs\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.025258 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-config-data\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.026608 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data-custom\") pod \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.026718 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92g7g\" (UniqueName: \"kubernetes.io/projected/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-kube-api-access-92g7g\") pod \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.026743 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-combined-ca-bundle\") pod \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\" (UID: \"23d4ece9-a8c0-4842-a3d3-d5c05c77676b\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.027049 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.027075 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-combined-ca-bundle\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.028388 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-logs" (OuterVolumeSpecName: "logs") pod "23d4ece9-a8c0-4842-a3d3-d5c05c77676b" (UID: "23d4ece9-a8c0-4842-a3d3-d5c05c77676b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.035085 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.040750 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "23d4ece9-a8c0-4842-a3d3-d5c05c77676b" (UID: "23d4ece9-a8c0-4842-a3d3-d5c05c77676b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.063908 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.100595 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.100713 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-logs" (OuterVolumeSpecName: "logs") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.128985 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.143194 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "23d4ece9-a8c0-4842-a3d3-d5c05c77676b" (UID: "23d4ece9-a8c0-4842-a3d3-d5c05c77676b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.171732 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv" (OuterVolumeSpecName: "kube-api-access-hdvbv") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "kube-api-access-hdvbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.178189 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179576 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts" (OuterVolumeSpecName: "scripts") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.179780 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="proxy-httpd" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179826 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="proxy-httpd" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.179854 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api-log" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179860 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api-log" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.179883 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-central-agent" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179890 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-central-agent" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.179906 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-log" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179912 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-log" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.179925 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179934 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.179952 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-httpd" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179958 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-httpd" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.179985 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-notification-agent" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.179991 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-notification-agent" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.180009 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.180016 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.180032 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="sg-core" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.180038 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="sg-core" Dec 06 10:57:21 crc kubenswrapper[4678]: E1206 10:57:21.180059 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.180066 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.181134 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdvbv\" (UniqueName: \"kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.181452 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts\") pod \"ccee4c53-a485-4390-b65f-c2f55f2881ad\" (UID: \"ccee4c53-a485-4390-b65f-c2f55f2881ad\") " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.182554 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.182573 4678 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.182588 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ccee4c53-a485-4390-b65f-c2f55f2881ad-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.182596 4678 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: W1206 10:57:21.182933 4678 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/ccee4c53-a485-4390-b65f-c2f55f2881ad/volumes/kubernetes.io~secret/scripts Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.182952 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts" (OuterVolumeSpecName: "scripts") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: W1206 10:57:21.183010 4678 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/ccee4c53-a485-4390-b65f-c2f55f2881ad/volumes/kubernetes.io~projected/kube-api-access-hdvbv Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.183058 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv" (OuterVolumeSpecName: "kube-api-access-hdvbv") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "kube-api-access-hdvbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191063 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-log" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191091 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api-log" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191109 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="proxy-httpd" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191131 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191147 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-notification-agent" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191155 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" containerName="glance-httpd" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191172 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="sg-core" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191186 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" containerName="ceilometer-central-agent" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191200 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" containerName="barbican-api" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.191220 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" containerName="cinder-api-log" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.202206 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-scripts" (OuterVolumeSpecName: "scripts") pod "23d4ece9-a8c0-4842-a3d3-d5c05c77676b" (UID: "23d4ece9-a8c0-4842-a3d3-d5c05c77676b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.212657 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.239764 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.263877 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-kube-api-access-92g7g" (OuterVolumeSpecName: "kube-api-access-92g7g") pod "23d4ece9-a8c0-4842-a3d3-d5c05c77676b" (UID: "23d4ece9-a8c0-4842-a3d3-d5c05c77676b"). InnerVolumeSpecName "kube-api-access-92g7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.264820 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.264965 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.287507 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdvbv\" (UniqueName: \"kubernetes.io/projected/ccee4c53-a485-4390-b65f-c2f55f2881ad-kube-api-access-hdvbv\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.287548 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92g7g\" (UniqueName: \"kubernetes.io/projected/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-kube-api-access-92g7g\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.287586 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.287599 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.287612 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.295542 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.389050 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-log-httpd\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.389092 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.389132 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-run-httpd\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.389147 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-scripts\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.389194 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clsxm\" (UniqueName: \"kubernetes.io/projected/67975c6d-b4e1-4497-a46d-4abe4927432c-kube-api-access-clsxm\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.389247 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.389269 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-config-data\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.423314 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.441035 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493585 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clsxm\" (UniqueName: \"kubernetes.io/projected/67975c6d-b4e1-4497-a46d-4abe4927432c-kube-api-access-clsxm\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493695 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493731 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-config-data\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493773 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-log-httpd\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493801 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493843 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-run-httpd\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493867 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-scripts\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493930 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.493944 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.496683 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-log-httpd\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.497849 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-run-httpd\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.514160 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.521986 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3899eb2-a276-4bc7-97d5-268ea406829a" path="/var/lib/kubelet/pods/a3899eb2-a276-4bc7-97d5-268ea406829a/volumes" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.525212 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd23d8eb-f06c-48e4-9319-aa699a260ca0" path="/var/lib/kubelet/pods/cd23d8eb-f06c-48e4-9319-aa699a260ca0/volumes" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.547819 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clsxm\" (UniqueName: \"kubernetes.io/projected/67975c6d-b4e1-4497-a46d-4abe4927432c-kube-api-access-clsxm\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.553101 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.554236 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "23d4ece9-a8c0-4842-a3d3-d5c05c77676b" (UID: "23d4ece9-a8c0-4842-a3d3-d5c05c77676b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.554444 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-config-data\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.556890 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-config-data" (OuterVolumeSpecName: "config-data") pod "ccee4c53-a485-4390-b65f-c2f55f2881ad" (UID: "ccee4c53-a485-4390-b65f-c2f55f2881ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.557416 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.560958 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-scripts\") pod \"ceilometer-0\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.596007 4678 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.596137 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccee4c53-a485-4390-b65f-c2f55f2881ad-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.596209 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.657330 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data" (OuterVolumeSpecName: "config-data") pod "23d4ece9-a8c0-4842-a3d3-d5c05c77676b" (UID: "23d4ece9-a8c0-4842-a3d3-d5c05c77676b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.697858 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23d4ece9-a8c0-4842-a3d3-d5c05c77676b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.751047 4678 generic.go:334] "Generic (PLEG): container finished" podID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerID="7c08a0e53ba728e5e8bef5ac77f4984309f75903c6d92ea13330a8497fd15033" exitCode=0 Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.769190 4678 generic.go:334] "Generic (PLEG): container finished" podID="1c3dd118-3077-4c96-b13b-edb34a16d944" containerID="f226fac5ff92d5521c8126125c7ea7b1f138f5f0939b61fbd9643a0c40b5741f" exitCode=0 Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.778298 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-7391-account-create-update-74ks6" podStartSLOduration=4.778286387 podStartE2EDuration="4.778286387s" podCreationTimestamp="2025-12-06 10:57:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:21.775910602 +0000 UTC m=+1246.619342041" watchObservedRunningTime="2025-12-06 10:57:21.778286387 +0000 UTC m=+1246.621717826" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.783394 4678 generic.go:334] "Generic (PLEG): container finished" podID="a54c6d51-0b8e-491d-b5fb-8d08a260232a" containerID="913a209132c2adb82e0be8f3c4fd4db2fefc66393db79043d32d41e49e286137" exitCode=0 Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.785101 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.794052 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.833089 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-4zkvd" podStartSLOduration=4.833066234 podStartE2EDuration="4.833066234s" podCreationTimestamp="2025-12-06 10:57:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:21.829610045 +0000 UTC m=+1246.673041484" watchObservedRunningTime="2025-12-06 10:57:21.833066234 +0000 UTC m=+1246.676497673" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.853287 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-8kxvz" podStartSLOduration=4.853268831 podStartE2EDuration="4.853268831s" podCreationTimestamp="2025-12-06 10:57:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:21.848448259 +0000 UTC m=+1246.691879698" watchObservedRunningTime="2025-12-06 10:57:21.853268831 +0000 UTC m=+1246.696700270" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.865853 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9aac132-0acd-48d8-a2e7-2d3b82157677","Type":"ContainerDied","Data":"7c08a0e53ba728e5e8bef5ac77f4984309f75903c6d92ea13330a8497fd15033"} Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.867562 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7391-account-create-update-74ks6" event={"ID":"582eb659-cea1-44f9-9d87-a2d965a9bcfc","Type":"ContainerStarted","Data":"7024ffc23101ee98af7e162435be201e29c64be1737c12844284f2d9ed2950f6"} Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.867604 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4zkvd" event={"ID":"d549492c-fcf7-402b-95e5-c0bee52f86eb","Type":"ContainerStarted","Data":"2fd015cd365d86d0f36cd85f92c016d03e9866baea84532882fda84116b71284"} Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.867616 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" event={"ID":"1c3dd118-3077-4c96-b13b-edb34a16d944","Type":"ContainerDied","Data":"f226fac5ff92d5521c8126125c7ea7b1f138f5f0939b61fbd9643a0c40b5741f"} Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.867728 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8kxvz" event={"ID":"081454ce-456d-4cd3-bec3-b7a06034ab39","Type":"ContainerStarted","Data":"7c02b4470a48fb0bbae65eb0738605e21260798f2505835287da1c0d656eca4e"} Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.867743 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6btts" event={"ID":"a54c6d51-0b8e-491d-b5fb-8d08a260232a","Type":"ContainerDied","Data":"913a209132c2adb82e0be8f3c4fd4db2fefc66393db79043d32d41e49e286137"} Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.867761 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-455a-account-create-update-kg2j5" event={"ID":"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee","Type":"ContainerStarted","Data":"bdfaabcfa35fb516f840b0a76fcd3749e0e5ec8ddfd8f9a6ef6825000f7bdc3b"} Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.914449 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.952358 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-455a-account-create-update-kg2j5" podStartSLOduration=4.952340875 podStartE2EDuration="4.952340875s" podCreationTimestamp="2025-12-06 10:57:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:21.904576959 +0000 UTC m=+1246.748008398" watchObservedRunningTime="2025-12-06 10:57:21.952340875 +0000 UTC m=+1246.795772314" Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.975393 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:57:21 crc kubenswrapper[4678]: I1206 10:57:21.987091 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.023002 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.045918 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.069351 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.072426 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.078134 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.078937 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.079047 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.096204 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.116656 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.118698 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.125567 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.125863 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.161351 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.217845 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-config-data\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.217908 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b283a032-2cad-40a7-bb76-88a4f3f47472-logs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.217948 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-config-data-custom\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.217985 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218015 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218040 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-scripts\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218073 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72g4r\" (UniqueName: \"kubernetes.io/projected/ef3105b5-af55-4aec-99a8-10792740d742-kube-api-access-72g4r\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218091 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218113 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef3105b5-af55-4aec-99a8-10792740d742-logs\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218145 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218162 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218193 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm6b6\" (UniqueName: \"kubernetes.io/projected/b283a032-2cad-40a7-bb76-88a4f3f47472-kube-api-access-dm6b6\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218229 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218263 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b283a032-2cad-40a7-bb76-88a4f3f47472-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218328 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef3105b5-af55-4aec-99a8-10792740d742-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218353 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.218406 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321473 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-config-data\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321525 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b283a032-2cad-40a7-bb76-88a4f3f47472-logs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321565 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-config-data-custom\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321587 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321617 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321638 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-scripts\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321653 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72g4r\" (UniqueName: \"kubernetes.io/projected/ef3105b5-af55-4aec-99a8-10792740d742-kube-api-access-72g4r\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321701 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef3105b5-af55-4aec-99a8-10792740d742-logs\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321723 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321741 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321768 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm6b6\" (UniqueName: \"kubernetes.io/projected/b283a032-2cad-40a7-bb76-88a4f3f47472-kube-api-access-dm6b6\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321789 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321814 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b283a032-2cad-40a7-bb76-88a4f3f47472-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321843 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef3105b5-af55-4aec-99a8-10792740d742-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321862 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.321885 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.324135 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ef3105b5-af55-4aec-99a8-10792740d742-logs\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.324221 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b283a032-2cad-40a7-bb76-88a4f3f47472-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.324923 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.325675 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ef3105b5-af55-4aec-99a8-10792740d742-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.326932 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.337140 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b283a032-2cad-40a7-bb76-88a4f3f47472-logs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.342369 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-scripts\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.342848 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.343415 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.345111 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.345282 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.347275 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-scripts\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.347886 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-config-data-custom\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.351030 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-config-data\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.354819 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef3105b5-af55-4aec-99a8-10792740d742-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.380262 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b283a032-2cad-40a7-bb76-88a4f3f47472-config-data\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.400888 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm6b6\" (UniqueName: \"kubernetes.io/projected/b283a032-2cad-40a7-bb76-88a4f3f47472-kube-api-access-dm6b6\") pod \"cinder-api-0\" (UID: \"b283a032-2cad-40a7-bb76-88a4f3f47472\") " pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.423988 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-logs\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.424126 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-scripts\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.424162 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.424207 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzhwt\" (UniqueName: \"kubernetes.io/projected/d9aac132-0acd-48d8-a2e7-2d3b82157677-kube-api-access-lzhwt\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.424242 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-httpd-run\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.424285 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-internal-tls-certs\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.424315 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-combined-ca-bundle\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.424353 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-config-data\") pod \"d9aac132-0acd-48d8-a2e7-2d3b82157677\" (UID: \"d9aac132-0acd-48d8-a2e7-2d3b82157677\") " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.427722 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72g4r\" (UniqueName: \"kubernetes.io/projected/ef3105b5-af55-4aec-99a8-10792740d742-kube-api-access-72g4r\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.430907 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-logs" (OuterVolumeSpecName: "logs") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.432428 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.456703 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.457735 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-scripts" (OuterVolumeSpecName: "scripts") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.467224 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.468738 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9aac132-0acd-48d8-a2e7-2d3b82157677-kube-api-access-lzhwt" (OuterVolumeSpecName: "kube-api-access-lzhwt") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "kube-api-access-lzhwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.503001 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"ef3105b5-af55-4aec-99a8-10792740d742\") " pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.517352 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.546684 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.546720 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzhwt\" (UniqueName: \"kubernetes.io/projected/d9aac132-0acd-48d8-a2e7-2d3b82157677-kube-api-access-lzhwt\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.546749 4678 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.546762 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.546772 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9aac132-0acd-48d8-a2e7-2d3b82157677-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.546781 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.547804 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.575325 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.605080 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.654676 4678 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.654711 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.739276 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-config-data" (OuterVolumeSpecName: "config-data") pod "d9aac132-0acd-48d8-a2e7-2d3b82157677" (UID: "d9aac132-0acd-48d8-a2e7-2d3b82157677"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.760224 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9aac132-0acd-48d8-a2e7-2d3b82157677-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.811325 4678 generic.go:334] "Generic (PLEG): container finished" podID="582eb659-cea1-44f9-9d87-a2d965a9bcfc" containerID="7024ffc23101ee98af7e162435be201e29c64be1737c12844284f2d9ed2950f6" exitCode=0 Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.811379 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7391-account-create-update-74ks6" event={"ID":"582eb659-cea1-44f9-9d87-a2d965a9bcfc","Type":"ContainerDied","Data":"7024ffc23101ee98af7e162435be201e29c64be1737c12844284f2d9ed2950f6"} Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.811823 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.815091 4678 generic.go:334] "Generic (PLEG): container finished" podID="d549492c-fcf7-402b-95e5-c0bee52f86eb" containerID="2fd015cd365d86d0f36cd85f92c016d03e9866baea84532882fda84116b71284" exitCode=0 Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.815147 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4zkvd" event={"ID":"d549492c-fcf7-402b-95e5-c0bee52f86eb","Type":"ContainerDied","Data":"2fd015cd365d86d0f36cd85f92c016d03e9866baea84532882fda84116b71284"} Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.818506 4678 generic.go:334] "Generic (PLEG): container finished" podID="081454ce-456d-4cd3-bec3-b7a06034ab39" containerID="7c02b4470a48fb0bbae65eb0738605e21260798f2505835287da1c0d656eca4e" exitCode=0 Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.818545 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8kxvz" event={"ID":"081454ce-456d-4cd3-bec3-b7a06034ab39","Type":"ContainerDied","Data":"7c02b4470a48fb0bbae65eb0738605e21260798f2505835287da1c0d656eca4e"} Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.821089 4678 generic.go:334] "Generic (PLEG): container finished" podID="cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee" containerID="bdfaabcfa35fb516f840b0a76fcd3749e0e5ec8ddfd8f9a6ef6825000f7bdc3b" exitCode=0 Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.821125 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-455a-account-create-update-kg2j5" event={"ID":"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee","Type":"ContainerDied","Data":"bdfaabcfa35fb516f840b0a76fcd3749e0e5ec8ddfd8f9a6ef6825000f7bdc3b"} Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.838983 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-f5f494679-7bm9s" event={"ID":"b31f2128-db6e-498b-bd38-d0b1c41a4603","Type":"ContainerStarted","Data":"0cc3f42e8d08637e1638195fa7d1b126e3c93e7c20dde5f5f13bef6b179146cd"} Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.839837 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.839869 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.871189 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d9aac132-0acd-48d8-a2e7-2d3b82157677","Type":"ContainerDied","Data":"5ed2a46b5656e3e2ee06bda2750c2049628d2b2a1b04618bb2d8983f5838d523"} Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.871242 4678 scope.go:117] "RemoveContainer" containerID="7c08a0e53ba728e5e8bef5ac77f4984309f75903c6d92ea13330a8497fd15033" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.871294 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:22 crc kubenswrapper[4678]: W1206 10:57:22.878844 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67975c6d_b4e1_4497_a46d_4abe4927432c.slice/crio-a55eb53c016b68d35662ec8b5128d185c60b2c58fe266a5d845bd04ad3be7e8d WatchSource:0}: Error finding container a55eb53c016b68d35662ec8b5128d185c60b2c58fe266a5d845bd04ad3be7e8d: Status 404 returned error can't find the container with id a55eb53c016b68d35662ec8b5128d185c60b2c58fe266a5d845bd04ad3be7e8d Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.960591 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-f5f494679-7bm9s" podStartSLOduration=13.960577665 podStartE2EDuration="13.960577665s" podCreationTimestamp="2025-12-06 10:57:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:22.959811002 +0000 UTC m=+1247.803242441" watchObservedRunningTime="2025-12-06 10:57:22.960577665 +0000 UTC m=+1247.804009104" Dec 06 10:57:22 crc kubenswrapper[4678]: I1206 10:57:22.988658 4678 scope.go:117] "RemoveContainer" containerID="abe07ac3ca1ab61beacbed1af31ef4a382dcd3d073c4b53028ae90606ff58fc7" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.013129 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.053659 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.058908 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:57:23 crc kubenswrapper[4678]: E1206 10:57:23.059261 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-httpd" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.059281 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-httpd" Dec 06 10:57:23 crc kubenswrapper[4678]: E1206 10:57:23.059306 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-log" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.059315 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-log" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.059622 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-httpd" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.059651 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" containerName="glance-log" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.081972 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.082088 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.088456 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.094708 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.251599 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 06 10:57:23 crc kubenswrapper[4678]: W1206 10:57:23.264933 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb283a032_2cad_40a7_bb76_88a4f3f47472.slice/crio-9e4a7e3e46efe318cd2a48807fd48b2473eedcfc1090dcff582fb225160dbd07 WatchSource:0}: Error finding container 9e4a7e3e46efe318cd2a48807fd48b2473eedcfc1090dcff582fb225160dbd07: Status 404 returned error can't find the container with id 9e4a7e3e46efe318cd2a48807fd48b2473eedcfc1090dcff582fb225160dbd07 Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.289956 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965cdf20-a198-4cc8-9a8f-700388dbf4cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.290059 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghnfw\" (UniqueName: \"kubernetes.io/projected/965cdf20-a198-4cc8-9a8f-700388dbf4cd-kube-api-access-ghnfw\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.290205 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965cdf20-a198-4cc8-9a8f-700388dbf4cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.290252 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.290273 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.290306 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.290343 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.290374 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393528 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965cdf20-a198-4cc8-9a8f-700388dbf4cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393580 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393604 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393627 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393650 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393669 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393740 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965cdf20-a198-4cc8-9a8f-700388dbf4cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.393761 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghnfw\" (UniqueName: \"kubernetes.io/projected/965cdf20-a198-4cc8-9a8f-700388dbf4cd-kube-api-access-ghnfw\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.394454 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/965cdf20-a198-4cc8-9a8f-700388dbf4cd-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.395067 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/965cdf20-a198-4cc8-9a8f-700388dbf4cd-logs\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.395306 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.408319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-config-data\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.409804 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-scripts\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.410352 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.432235 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/965cdf20-a198-4cc8-9a8f-700388dbf4cd-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.434749 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghnfw\" (UniqueName: \"kubernetes.io/projected/965cdf20-a198-4cc8-9a8f-700388dbf4cd-kube-api-access-ghnfw\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.523202 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"965cdf20-a198-4cc8-9a8f-700388dbf4cd\") " pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.567559 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23d4ece9-a8c0-4842-a3d3-d5c05c77676b" path="/var/lib/kubelet/pods/23d4ece9-a8c0-4842-a3d3-d5c05c77676b/volumes" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.572755 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccee4c53-a485-4390-b65f-c2f55f2881ad" path="/var/lib/kubelet/pods/ccee4c53-a485-4390-b65f-c2f55f2881ad/volumes" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.574204 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9aac132-0acd-48d8-a2e7-2d3b82157677" path="/var/lib/kubelet/pods/d9aac132-0acd-48d8-a2e7-2d3b82157677/volumes" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.615106 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.636922 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.688498 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.704373 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.706419 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk8hh\" (UniqueName: \"kubernetes.io/projected/1c3dd118-3077-4c96-b13b-edb34a16d944-kube-api-access-tk8hh\") pod \"1c3dd118-3077-4c96-b13b-edb34a16d944\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.711719 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a54c6d51-0b8e-491d-b5fb-8d08a260232a-operator-scripts\") pod \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.711790 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3dd118-3077-4c96-b13b-edb34a16d944-operator-scripts\") pod \"1c3dd118-3077-4c96-b13b-edb34a16d944\" (UID: \"1c3dd118-3077-4c96-b13b-edb34a16d944\") " Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.711829 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvtmr\" (UniqueName: \"kubernetes.io/projected/a54c6d51-0b8e-491d-b5fb-8d08a260232a-kube-api-access-pvtmr\") pod \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\" (UID: \"a54c6d51-0b8e-491d-b5fb-8d08a260232a\") " Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.713625 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c3dd118-3077-4c96-b13b-edb34a16d944-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c3dd118-3077-4c96-b13b-edb34a16d944" (UID: "1c3dd118-3077-4c96-b13b-edb34a16d944"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.714158 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c3dd118-3077-4c96-b13b-edb34a16d944-kube-api-access-tk8hh" (OuterVolumeSpecName: "kube-api-access-tk8hh") pod "1c3dd118-3077-4c96-b13b-edb34a16d944" (UID: "1c3dd118-3077-4c96-b13b-edb34a16d944"). InnerVolumeSpecName "kube-api-access-tk8hh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.722078 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a54c6d51-0b8e-491d-b5fb-8d08a260232a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a54c6d51-0b8e-491d-b5fb-8d08a260232a" (UID: "a54c6d51-0b8e-491d-b5fb-8d08a260232a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.723675 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a54c6d51-0b8e-491d-b5fb-8d08a260232a-kube-api-access-pvtmr" (OuterVolumeSpecName: "kube-api-access-pvtmr") pod "a54c6d51-0b8e-491d-b5fb-8d08a260232a" (UID: "a54c6d51-0b8e-491d-b5fb-8d08a260232a"). InnerVolumeSpecName "kube-api-access-pvtmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.814632 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a54c6d51-0b8e-491d-b5fb-8d08a260232a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.814667 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3dd118-3077-4c96-b13b-edb34a16d944-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.814676 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvtmr\" (UniqueName: \"kubernetes.io/projected/a54c6d51-0b8e-491d-b5fb-8d08a260232a-kube-api-access-pvtmr\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.814685 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk8hh\" (UniqueName: \"kubernetes.io/projected/1c3dd118-3077-4c96-b13b-edb34a16d944-kube-api-access-tk8hh\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.897326 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b283a032-2cad-40a7-bb76-88a4f3f47472","Type":"ContainerStarted","Data":"9e4a7e3e46efe318cd2a48807fd48b2473eedcfc1090dcff582fb225160dbd07"} Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.905958 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerStarted","Data":"a55eb53c016b68d35662ec8b5128d185c60b2c58fe266a5d845bd04ad3be7e8d"} Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.907754 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef3105b5-af55-4aec-99a8-10792740d742","Type":"ContainerStarted","Data":"f51fbceeba1b507610bb1a6e14b8bc206d95ed081c7d371f62e0b093dc640af0"} Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.912963 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" event={"ID":"1c3dd118-3077-4c96-b13b-edb34a16d944","Type":"ContainerDied","Data":"5ec3851341dcce486cfcfc56f923a3abef4f7681a830c5fd806c3698c9798cb6"} Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.912991 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ec3851341dcce486cfcfc56f923a3abef4f7681a830c5fd806c3698c9798cb6" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.913014 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-39d4-account-create-update-wpqkc" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.917092 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6btts" event={"ID":"a54c6d51-0b8e-491d-b5fb-8d08a260232a","Type":"ContainerDied","Data":"f99befe25f574d09563702d36a8f61a3748ddac5be2961cafdc2933de2220144"} Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.917133 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f99befe25f574d09563702d36a8f61a3748ddac5be2961cafdc2933de2220144" Dec 06 10:57:23 crc kubenswrapper[4678]: I1206 10:57:23.917633 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6btts" Dec 06 10:57:24 crc kubenswrapper[4678]: I1206 10:57:24.989249 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-8kxvz" event={"ID":"081454ce-456d-4cd3-bec3-b7a06034ab39","Type":"ContainerDied","Data":"1f8bf9573b476c1a5bea2e6927ac5c9c809a8683ed3bc8646ac581a40eca774a"} Dec 06 10:57:24 crc kubenswrapper[4678]: I1206 10:57:24.989852 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f8bf9573b476c1a5bea2e6927ac5c9c809a8683ed3bc8646ac581a40eca774a" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.005826 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-455a-account-create-update-kg2j5" event={"ID":"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee","Type":"ContainerDied","Data":"be6ae9188bdb56aac9b830bcc6c1c51c5ab351ae2fee1236c3aaca1e6cf12c2c"} Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.006097 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be6ae9188bdb56aac9b830bcc6c1c51c5ab351ae2fee1236c3aaca1e6cf12c2c" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.018523 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.025441 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerStarted","Data":"462206c286d6b9599b99e88697841141402368deba621ee5f451284246860e39"} Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.031837 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7391-account-create-update-74ks6" event={"ID":"582eb659-cea1-44f9-9d87-a2d965a9bcfc","Type":"ContainerDied","Data":"e02b7b915e3719a8f8f0f94bd71960314bb72332ad28120c0a31098157bb9bed"} Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.031875 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e02b7b915e3719a8f8f0f94bd71960314bb72332ad28120c0a31098157bb9bed" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.039566 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-4zkvd" event={"ID":"d549492c-fcf7-402b-95e5-c0bee52f86eb","Type":"ContainerDied","Data":"6289bfa16d8e28be70e854c055f704746abd207cfd7a3d721df59face8a5b3c9"} Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.039619 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6289bfa16d8e28be70e854c055f704746abd207cfd7a3d721df59face8a5b3c9" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.055309 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.089605 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmkfh\" (UniqueName: \"kubernetes.io/projected/081454ce-456d-4cd3-bec3-b7a06034ab39-kube-api-access-xmkfh\") pod \"081454ce-456d-4cd3-bec3-b7a06034ab39\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.089758 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081454ce-456d-4cd3-bec3-b7a06034ab39-operator-scripts\") pod \"081454ce-456d-4cd3-bec3-b7a06034ab39\" (UID: \"081454ce-456d-4cd3-bec3-b7a06034ab39\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.092145 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/081454ce-456d-4cd3-bec3-b7a06034ab39-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "081454ce-456d-4cd3-bec3-b7a06034ab39" (UID: "081454ce-456d-4cd3-bec3-b7a06034ab39"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.094386 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.103001 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-f5f494679-7bm9s" podUID="b31f2128-db6e-498b-bd38-d0b1c41a4603" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.103398 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.142293 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/081454ce-456d-4cd3-bec3-b7a06034ab39-kube-api-access-xmkfh" (OuterVolumeSpecName: "kube-api-access-xmkfh") pod "081454ce-456d-4cd3-bec3-b7a06034ab39" (UID: "081454ce-456d-4cd3-bec3-b7a06034ab39"). InnerVolumeSpecName "kube-api-access-xmkfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.154402 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.190752 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582eb659-cea1-44f9-9d87-a2d965a9bcfc-operator-scripts\") pod \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.190917 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-operator-scripts\") pod \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.190970 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55wsd\" (UniqueName: \"kubernetes.io/projected/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-kube-api-access-55wsd\") pod \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\" (UID: \"cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.191043 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bd7n\" (UniqueName: \"kubernetes.io/projected/582eb659-cea1-44f9-9d87-a2d965a9bcfc-kube-api-access-5bd7n\") pod \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\" (UID: \"582eb659-cea1-44f9-9d87-a2d965a9bcfc\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.191343 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/582eb659-cea1-44f9-9d87-a2d965a9bcfc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "582eb659-cea1-44f9-9d87-a2d965a9bcfc" (UID: "582eb659-cea1-44f9-9d87-a2d965a9bcfc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.191426 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmkfh\" (UniqueName: \"kubernetes.io/projected/081454ce-456d-4cd3-bec3-b7a06034ab39-kube-api-access-xmkfh\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.191438 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/582eb659-cea1-44f9-9d87-a2d965a9bcfc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.191446 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/081454ce-456d-4cd3-bec3-b7a06034ab39-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.191675 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee" (UID: "cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.215633 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/582eb659-cea1-44f9-9d87-a2d965a9bcfc-kube-api-access-5bd7n" (OuterVolumeSpecName: "kube-api-access-5bd7n") pod "582eb659-cea1-44f9-9d87-a2d965a9bcfc" (UID: "582eb659-cea1-44f9-9d87-a2d965a9bcfc"). InnerVolumeSpecName "kube-api-access-5bd7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.225048 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-kube-api-access-55wsd" (OuterVolumeSpecName: "kube-api-access-55wsd") pod "cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee" (UID: "cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee"). InnerVolumeSpecName "kube-api-access-55wsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.293944 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stgp2\" (UniqueName: \"kubernetes.io/projected/d549492c-fcf7-402b-95e5-c0bee52f86eb-kube-api-access-stgp2\") pod \"d549492c-fcf7-402b-95e5-c0bee52f86eb\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.294069 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d549492c-fcf7-402b-95e5-c0bee52f86eb-operator-scripts\") pod \"d549492c-fcf7-402b-95e5-c0bee52f86eb\" (UID: \"d549492c-fcf7-402b-95e5-c0bee52f86eb\") " Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.294508 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55wsd\" (UniqueName: \"kubernetes.io/projected/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-kube-api-access-55wsd\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.294519 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bd7n\" (UniqueName: \"kubernetes.io/projected/582eb659-cea1-44f9-9d87-a2d965a9bcfc-kube-api-access-5bd7n\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.294528 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.294889 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d549492c-fcf7-402b-95e5-c0bee52f86eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d549492c-fcf7-402b-95e5-c0bee52f86eb" (UID: "d549492c-fcf7-402b-95e5-c0bee52f86eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.358738 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d549492c-fcf7-402b-95e5-c0bee52f86eb-kube-api-access-stgp2" (OuterVolumeSpecName: "kube-api-access-stgp2") pod "d549492c-fcf7-402b-95e5-c0bee52f86eb" (UID: "d549492c-fcf7-402b-95e5-c0bee52f86eb"). InnerVolumeSpecName "kube-api-access-stgp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.396850 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stgp2\" (UniqueName: \"kubernetes.io/projected/d549492c-fcf7-402b-95e5-c0bee52f86eb-kube-api-access-stgp2\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:25 crc kubenswrapper[4678]: I1206 10:57:25.396882 4678 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d549492c-fcf7-402b-95e5-c0bee52f86eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.073029 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerStarted","Data":"e9d8ef3b9b3197c39a7431826b493d8f9f098825e8accecbc18ba13ebe781500"} Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.093297 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef3105b5-af55-4aec-99a8-10792740d742","Type":"ContainerStarted","Data":"424d8a8a15d6c83548ac3e2ebd1a1afc5dd33cce7e76451b92d22c084677890e"} Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.099179 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965cdf20-a198-4cc8-9a8f-700388dbf4cd","Type":"ContainerStarted","Data":"f6a92c6348182ac8e7cdaf0c88dfe20cab64f511906510632dbc3670af3d0b8f"} Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.110291 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-455a-account-create-update-kg2j5" Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.110840 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-8kxvz" Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.110864 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7391-account-create-update-74ks6" Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.110905 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b283a032-2cad-40a7-bb76-88a4f3f47472","Type":"ContainerStarted","Data":"b8604b086e9913cc01881f4918140c9851dcd5beef707d23dd3034c454830348"} Dec 06 10:57:26 crc kubenswrapper[4678]: I1206 10:57:26.110945 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-4zkvd" Dec 06 10:57:27 crc kubenswrapper[4678]: I1206 10:57:27.129186 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"ef3105b5-af55-4aec-99a8-10792740d742","Type":"ContainerStarted","Data":"26474094b617a9dcdff3dceeae9b4bc49ac7452885bf7586a0ced369cd9b2ca3"} Dec 06 10:57:27 crc kubenswrapper[4678]: I1206 10:57:27.137588 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965cdf20-a198-4cc8-9a8f-700388dbf4cd","Type":"ContainerStarted","Data":"584465c727c353bf2958b38c15b7677c6ee46f01615201612937f632e1d74aed"} Dec 06 10:57:27 crc kubenswrapper[4678]: I1206 10:57:27.140835 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b283a032-2cad-40a7-bb76-88a4f3f47472","Type":"ContainerStarted","Data":"e76699bc30b06139f2440f47b09e1a24407ed3e90d10746c187149bcc766c1b1"} Dec 06 10:57:27 crc kubenswrapper[4678]: I1206 10:57:27.141522 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 06 10:57:27 crc kubenswrapper[4678]: I1206 10:57:27.144851 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerStarted","Data":"1b6d08ac0901c584f1754ede48f1185546240b9f2293b23caebc2672ff2cc0f6"} Dec 06 10:57:27 crc kubenswrapper[4678]: I1206 10:57:27.202955 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.20294039 podStartE2EDuration="6.20294039s" podCreationTimestamp="2025-12-06 10:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:27.182805465 +0000 UTC m=+1252.026236904" watchObservedRunningTime="2025-12-06 10:57:27.20294039 +0000 UTC m=+1252.046371819" Dec 06 10:57:27 crc kubenswrapper[4678]: I1206 10:57:27.203623 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.203618842 podStartE2EDuration="6.203618842s" podCreationTimestamp="2025-12-06 10:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:27.200514174 +0000 UTC m=+1252.043945613" watchObservedRunningTime="2025-12-06 10:57:27.203618842 +0000 UTC m=+1252.047050281" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.013817 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdm9z"] Dec 06 10:57:28 crc kubenswrapper[4678]: E1206 10:57:28.014629 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="582eb659-cea1-44f9-9d87-a2d965a9bcfc" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.014724 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="582eb659-cea1-44f9-9d87-a2d965a9bcfc" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: E1206 10:57:28.014794 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3dd118-3077-4c96-b13b-edb34a16d944" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.014846 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3dd118-3077-4c96-b13b-edb34a16d944" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: E1206 10:57:28.014904 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="081454ce-456d-4cd3-bec3-b7a06034ab39" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.014961 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="081454ce-456d-4cd3-bec3-b7a06034ab39" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: E1206 10:57:28.015021 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015072 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: E1206 10:57:28.015146 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d549492c-fcf7-402b-95e5-c0bee52f86eb" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015201 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d549492c-fcf7-402b-95e5-c0bee52f86eb" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: E1206 10:57:28.015273 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a54c6d51-0b8e-491d-b5fb-8d08a260232a" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015332 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a54c6d51-0b8e-491d-b5fb-8d08a260232a" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015588 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c3dd118-3077-4c96-b13b-edb34a16d944" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015661 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a54c6d51-0b8e-491d-b5fb-8d08a260232a" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015720 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d549492c-fcf7-402b-95e5-c0bee52f86eb" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015782 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015844 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="582eb659-cea1-44f9-9d87-a2d965a9bcfc" containerName="mariadb-account-create-update" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.015913 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="081454ce-456d-4cd3-bec3-b7a06034ab39" containerName="mariadb-database-create" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.016525 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.019546 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.019687 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-t9whj" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.019848 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.029705 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdm9z"] Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.073248 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxz48\" (UniqueName: \"kubernetes.io/projected/4d1feb10-f542-4553-89fc-373d1940ac22-kube-api-access-sxz48\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.073307 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-scripts\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.073341 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-config-data\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.073377 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.170015 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerStarted","Data":"77c740a839d987f139735e792bc219edcccad1d96437a18057ca0091310d6d56"} Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.170166 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.174273 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxz48\" (UniqueName: \"kubernetes.io/projected/4d1feb10-f542-4553-89fc-373d1940ac22-kube-api-access-sxz48\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.174324 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-scripts\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.174363 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-config-data\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.174394 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.182420 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.185986 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-scripts\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.187875 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"965cdf20-a198-4cc8-9a8f-700388dbf4cd","Type":"ContainerStarted","Data":"9d9434b2cb6333ccde356e223749f5e668e62c58264651cc67e449238ff08814"} Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.204265 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-config-data\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.213715 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxz48\" (UniqueName: \"kubernetes.io/projected/4d1feb10-f542-4553-89fc-373d1940ac22-kube-api-access-sxz48\") pod \"nova-cell0-conductor-db-sync-xdm9z\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.245542 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.614005869 podStartE2EDuration="8.245522814s" podCreationTimestamp="2025-12-06 10:57:20 +0000 UTC" firstStartedPulling="2025-12-06 10:57:22.89474051 +0000 UTC m=+1247.738171949" lastFinishedPulling="2025-12-06 10:57:27.526257455 +0000 UTC m=+1252.369688894" observedRunningTime="2025-12-06 10:57:28.237193871 +0000 UTC m=+1253.080625310" watchObservedRunningTime="2025-12-06 10:57:28.245522814 +0000 UTC m=+1253.088954263" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.272562 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.272545656 podStartE2EDuration="5.272545656s" podCreationTimestamp="2025-12-06 10:57:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:57:28.266863787 +0000 UTC m=+1253.110295246" watchObservedRunningTime="2025-12-06 10:57:28.272545656 +0000 UTC m=+1253.115977085" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.333173 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.853642 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdm9z"] Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.976072 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:57:28 crc kubenswrapper[4678]: I1206 10:57:28.976178 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:57:29 crc kubenswrapper[4678]: I1206 10:57:29.200160 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" event={"ID":"4d1feb10-f542-4553-89fc-373d1940ac22","Type":"ContainerStarted","Data":"787bd25dc6d4f8709128e940721c0fa0ba0a5f5b1811f75b6325a9a6309b4ccb"} Dec 06 10:57:29 crc kubenswrapper[4678]: I1206 10:57:29.526342 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:29 crc kubenswrapper[4678]: I1206 10:57:29.527818 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-f5f494679-7bm9s" Dec 06 10:57:30 crc kubenswrapper[4678]: I1206 10:57:30.255805 4678 generic.go:334] "Generic (PLEG): container finished" podID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" containerID="a04f288a96a3e3ff2987759b3e148e1cbbffe2e84ab32bb6357b33ca748488f8" exitCode=137 Dec 06 10:57:30 crc kubenswrapper[4678]: I1206 10:57:30.257307 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d77bcbbf8-cfc4l" event={"ID":"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e","Type":"ContainerDied","Data":"a04f288a96a3e3ff2987759b3e148e1cbbffe2e84ab32bb6357b33ca748488f8"} Dec 06 10:57:30 crc kubenswrapper[4678]: I1206 10:57:30.257332 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d77bcbbf8-cfc4l" event={"ID":"c2ceb969-bd57-4345-840c-4cd8b2b7ca8e","Type":"ContainerStarted","Data":"1acdb0aeb9dbae01cead6f93031c1de2df4af4a4b24c63776ae1fa495c7130b8"} Dec 06 10:57:32 crc kubenswrapper[4678]: I1206 10:57:32.576723 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 10:57:32 crc kubenswrapper[4678]: I1206 10:57:32.577065 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 06 10:57:32 crc kubenswrapper[4678]: I1206 10:57:32.634777 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 10:57:32 crc kubenswrapper[4678]: I1206 10:57:32.636121 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 06 10:57:33 crc kubenswrapper[4678]: I1206 10:57:33.282628 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 10:57:33 crc kubenswrapper[4678]: I1206 10:57:33.282683 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 06 10:57:33 crc kubenswrapper[4678]: I1206 10:57:33.705898 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:33 crc kubenswrapper[4678]: I1206 10:57:33.706209 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:33 crc kubenswrapper[4678]: I1206 10:57:33.763702 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:33 crc kubenswrapper[4678]: I1206 10:57:33.767839 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:34 crc kubenswrapper[4678]: I1206 10:57:34.292120 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:34 crc kubenswrapper[4678]: I1206 10:57:34.292175 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:35 crc kubenswrapper[4678]: I1206 10:57:35.891059 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 06 10:57:36 crc kubenswrapper[4678]: I1206 10:57:36.320196 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:57:36 crc kubenswrapper[4678]: I1206 10:57:36.320432 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:57:37 crc kubenswrapper[4678]: I1206 10:57:37.293649 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:37 crc kubenswrapper[4678]: I1206 10:57:37.293920 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-central-agent" containerID="cri-o://462206c286d6b9599b99e88697841141402368deba621ee5f451284246860e39" gracePeriod=30 Dec 06 10:57:37 crc kubenswrapper[4678]: I1206 10:57:37.293981 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="proxy-httpd" containerID="cri-o://77c740a839d987f139735e792bc219edcccad1d96437a18057ca0091310d6d56" gracePeriod=30 Dec 06 10:57:37 crc kubenswrapper[4678]: I1206 10:57:37.294031 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="sg-core" containerID="cri-o://1b6d08ac0901c584f1754ede48f1185546240b9f2293b23caebc2672ff2cc0f6" gracePeriod=30 Dec 06 10:57:37 crc kubenswrapper[4678]: I1206 10:57:37.294069 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-notification-agent" containerID="cri-o://e9d8ef3b9b3197c39a7431826b493d8f9f098825e8accecbc18ba13ebe781500" gracePeriod=30 Dec 06 10:57:37 crc kubenswrapper[4678]: I1206 10:57:37.319533 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.173:3000/\": EOF" Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.344728 4678 generic.go:334] "Generic (PLEG): container finished" podID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerID="77c740a839d987f139735e792bc219edcccad1d96437a18057ca0091310d6d56" exitCode=0 Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.344939 4678 generic.go:334] "Generic (PLEG): container finished" podID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerID="1b6d08ac0901c584f1754ede48f1185546240b9f2293b23caebc2672ff2cc0f6" exitCode=2 Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.344947 4678 generic.go:334] "Generic (PLEG): container finished" podID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerID="462206c286d6b9599b99e88697841141402368deba621ee5f451284246860e39" exitCode=0 Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.344965 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerDied","Data":"77c740a839d987f139735e792bc219edcccad1d96437a18057ca0091310d6d56"} Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.344995 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerDied","Data":"1b6d08ac0901c584f1754ede48f1185546240b9f2293b23caebc2672ff2cc0f6"} Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.345009 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerDied","Data":"462206c286d6b9599b99e88697841141402368deba621ee5f451284246860e39"} Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.670192 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.670266 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.866056 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.866501 4678 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.920572 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 06 10:57:38 crc kubenswrapper[4678]: I1206 10:57:38.989579 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:57:39 crc kubenswrapper[4678]: I1206 10:57:39.016646 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 06 10:57:39 crc kubenswrapper[4678]: I1206 10:57:39.310404 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:57:39 crc kubenswrapper[4678]: I1206 10:57:39.312996 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:57:39 crc kubenswrapper[4678]: I1206 10:57:39.314436 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 10:57:39 crc kubenswrapper[4678]: I1206 10:57:39.421697 4678 generic.go:334] "Generic (PLEG): container finished" podID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerID="e9d8ef3b9b3197c39a7431826b493d8f9f098825e8accecbc18ba13ebe781500" exitCode=0 Dec 06 10:57:39 crc kubenswrapper[4678]: I1206 10:57:39.422629 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerDied","Data":"e9d8ef3b9b3197c39a7431826b493d8f9f098825e8accecbc18ba13ebe781500"} Dec 06 10:57:43 crc kubenswrapper[4678]: E1206 10:57:43.720600 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified" Dec 06 10:57:43 crc kubenswrapper[4678]: E1206 10:57:43.721350 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sxz48,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-xdm9z_openstack(4d1feb10-f542-4553-89fc-373d1940ac22): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 10:57:43 crc kubenswrapper[4678]: E1206 10:57:43.723285 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" podUID="4d1feb10-f542-4553-89fc-373d1940ac22" Dec 06 10:57:43 crc kubenswrapper[4678]: I1206 10:57:43.832758 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.005738 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-combined-ca-bundle\") pod \"67975c6d-b4e1-4497-a46d-4abe4927432c\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.005825 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-scripts\") pod \"67975c6d-b4e1-4497-a46d-4abe4927432c\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.005845 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-sg-core-conf-yaml\") pod \"67975c6d-b4e1-4497-a46d-4abe4927432c\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.005939 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-config-data\") pod \"67975c6d-b4e1-4497-a46d-4abe4927432c\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.005965 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-run-httpd\") pod \"67975c6d-b4e1-4497-a46d-4abe4927432c\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.006028 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clsxm\" (UniqueName: \"kubernetes.io/projected/67975c6d-b4e1-4497-a46d-4abe4927432c-kube-api-access-clsxm\") pod \"67975c6d-b4e1-4497-a46d-4abe4927432c\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.006082 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-log-httpd\") pod \"67975c6d-b4e1-4497-a46d-4abe4927432c\" (UID: \"67975c6d-b4e1-4497-a46d-4abe4927432c\") " Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.006867 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "67975c6d-b4e1-4497-a46d-4abe4927432c" (UID: "67975c6d-b4e1-4497-a46d-4abe4927432c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.007217 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "67975c6d-b4e1-4497-a46d-4abe4927432c" (UID: "67975c6d-b4e1-4497-a46d-4abe4927432c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.012396 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67975c6d-b4e1-4497-a46d-4abe4927432c-kube-api-access-clsxm" (OuterVolumeSpecName: "kube-api-access-clsxm") pod "67975c6d-b4e1-4497-a46d-4abe4927432c" (UID: "67975c6d-b4e1-4497-a46d-4abe4927432c"). InnerVolumeSpecName "kube-api-access-clsxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.022535 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-scripts" (OuterVolumeSpecName: "scripts") pod "67975c6d-b4e1-4497-a46d-4abe4927432c" (UID: "67975c6d-b4e1-4497-a46d-4abe4927432c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.046699 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "67975c6d-b4e1-4497-a46d-4abe4927432c" (UID: "67975c6d-b4e1-4497-a46d-4abe4927432c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.108620 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.108815 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clsxm\" (UniqueName: \"kubernetes.io/projected/67975c6d-b4e1-4497-a46d-4abe4927432c-kube-api-access-clsxm\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.108873 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/67975c6d-b4e1-4497-a46d-4abe4927432c-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.108985 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.109040 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.195644 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67975c6d-b4e1-4497-a46d-4abe4927432c" (UID: "67975c6d-b4e1-4497-a46d-4abe4927432c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.202653 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-config-data" (OuterVolumeSpecName: "config-data") pod "67975c6d-b4e1-4497-a46d-4abe4927432c" (UID: "67975c6d-b4e1-4497-a46d-4abe4927432c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.210782 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.210811 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67975c6d-b4e1-4497-a46d-4abe4927432c-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.488594 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"67975c6d-b4e1-4497-a46d-4abe4927432c","Type":"ContainerDied","Data":"a55eb53c016b68d35662ec8b5128d185c60b2c58fe266a5d845bd04ad3be7e8d"} Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.488840 4678 scope.go:117] "RemoveContainer" containerID="77c740a839d987f139735e792bc219edcccad1d96437a18057ca0091310d6d56" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.488621 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.521282 4678 scope.go:117] "RemoveContainer" containerID="1b6d08ac0901c584f1754ede48f1185546240b9f2293b23caebc2672ff2cc0f6" Dec 06 10:57:44 crc kubenswrapper[4678]: E1206 10:57:44.521556 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" podUID="4d1feb10-f542-4553-89fc-373d1940ac22" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.558266 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.571947 4678 scope.go:117] "RemoveContainer" containerID="e9d8ef3b9b3197c39a7431826b493d8f9f098825e8accecbc18ba13ebe781500" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.576097 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.586145 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:44 crc kubenswrapper[4678]: E1206 10:57:44.586796 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="sg-core" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.586822 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="sg-core" Dec 06 10:57:44 crc kubenswrapper[4678]: E1206 10:57:44.586848 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-notification-agent" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.586857 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-notification-agent" Dec 06 10:57:44 crc kubenswrapper[4678]: E1206 10:57:44.586878 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="proxy-httpd" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.586889 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="proxy-httpd" Dec 06 10:57:44 crc kubenswrapper[4678]: E1206 10:57:44.586905 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-central-agent" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.586913 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-central-agent" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.587153 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="sg-core" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.587175 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="proxy-httpd" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.587207 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-notification-agent" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.587230 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" containerName="ceilometer-central-agent" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.589301 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.591590 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.595726 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.600404 4678 scope.go:117] "RemoveContainer" containerID="462206c286d6b9599b99e88697841141402368deba621ee5f451284246860e39" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.649775 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.735899 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.735949 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.735969 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-scripts\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.736077 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-run-httpd\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.736098 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-log-httpd\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.736133 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfspt\" (UniqueName: \"kubernetes.io/projected/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-kube-api-access-hfspt\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.736164 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-config-data\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.838158 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-run-httpd\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.838208 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-log-httpd\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.838259 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfspt\" (UniqueName: \"kubernetes.io/projected/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-kube-api-access-hfspt\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.838306 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-config-data\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.838373 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.838759 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-log-httpd\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.839033 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.839055 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-run-httpd\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.839064 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-scripts\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.843220 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-config-data\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.843322 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.843369 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.844233 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-scripts\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.858169 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfspt\" (UniqueName: \"kubernetes.io/projected/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-kube-api-access-hfspt\") pod \"ceilometer-0\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " pod="openstack/ceilometer-0" Dec 06 10:57:44 crc kubenswrapper[4678]: I1206 10:57:44.915263 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:57:45 crc kubenswrapper[4678]: I1206 10:57:45.425291 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:45 crc kubenswrapper[4678]: I1206 10:57:45.488420 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67975c6d-b4e1-4497-a46d-4abe4927432c" path="/var/lib/kubelet/pods/67975c6d-b4e1-4497-a46d-4abe4927432c/volumes" Dec 06 10:57:45 crc kubenswrapper[4678]: I1206 10:57:45.499369 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerStarted","Data":"ece27320d31f716a330bcbef5153eacd6982fda566e8c0933702883a94ededeb"} Dec 06 10:57:46 crc kubenswrapper[4678]: I1206 10:57:46.509141 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerStarted","Data":"d28c28febe24024c8c3c243eddcc1589c9b2796739ed43f5f21f9b958f0647ed"} Dec 06 10:57:47 crc kubenswrapper[4678]: I1206 10:57:47.529631 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerStarted","Data":"52a269ade91dc4ed8a5efc1327192ce40516eddd39b6add978fd52e91c077477"} Dec 06 10:57:48 crc kubenswrapper[4678]: I1206 10:57:48.540238 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerStarted","Data":"3164efb7ea39f4016de8c9b1404e49089e5677c33c7a668ebf9bf8e3e1e34d27"} Dec 06 10:57:48 crc kubenswrapper[4678]: I1206 10:57:48.977548 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:57:49 crc kubenswrapper[4678]: I1206 10:57:49.310281 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d77bcbbf8-cfc4l" podUID="c2ceb969-bd57-4345-840c-4cd8b2b7ca8e" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Dec 06 10:57:49 crc kubenswrapper[4678]: I1206 10:57:49.431186 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:57:49 crc kubenswrapper[4678]: I1206 10:57:49.548984 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerStarted","Data":"6cf498d456910aa38c6f4503596ca4195471c3ab9bc0ea3dd2f50620380ec674"} Dec 06 10:57:49 crc kubenswrapper[4678]: I1206 10:57:49.550857 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 10:57:49 crc kubenswrapper[4678]: I1206 10:57:49.583696 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.450439612 podStartE2EDuration="5.583674316s" podCreationTimestamp="2025-12-06 10:57:44 +0000 UTC" firstStartedPulling="2025-12-06 10:57:45.423177012 +0000 UTC m=+1270.266608451" lastFinishedPulling="2025-12-06 10:57:48.556411716 +0000 UTC m=+1273.399843155" observedRunningTime="2025-12-06 10:57:49.56795123 +0000 UTC m=+1274.411382679" watchObservedRunningTime="2025-12-06 10:57:49.583674316 +0000 UTC m=+1274.427105755" Dec 06 10:57:50 crc kubenswrapper[4678]: I1206 10:57:50.558639 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-central-agent" containerID="cri-o://d28c28febe24024c8c3c243eddcc1589c9b2796739ed43f5f21f9b958f0647ed" gracePeriod=30 Dec 06 10:57:50 crc kubenswrapper[4678]: I1206 10:57:50.559178 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="proxy-httpd" containerID="cri-o://6cf498d456910aa38c6f4503596ca4195471c3ab9bc0ea3dd2f50620380ec674" gracePeriod=30 Dec 06 10:57:50 crc kubenswrapper[4678]: I1206 10:57:50.559324 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="sg-core" containerID="cri-o://3164efb7ea39f4016de8c9b1404e49089e5677c33c7a668ebf9bf8e3e1e34d27" gracePeriod=30 Dec 06 10:57:50 crc kubenswrapper[4678]: I1206 10:57:50.559377 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-notification-agent" containerID="cri-o://52a269ade91dc4ed8a5efc1327192ce40516eddd39b6add978fd52e91c077477" gracePeriod=30 Dec 06 10:57:51 crc kubenswrapper[4678]: I1206 10:57:51.569819 4678 generic.go:334] "Generic (PLEG): container finished" podID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerID="6cf498d456910aa38c6f4503596ca4195471c3ab9bc0ea3dd2f50620380ec674" exitCode=0 Dec 06 10:57:51 crc kubenswrapper[4678]: I1206 10:57:51.570053 4678 generic.go:334] "Generic (PLEG): container finished" podID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerID="3164efb7ea39f4016de8c9b1404e49089e5677c33c7a668ebf9bf8e3e1e34d27" exitCode=2 Dec 06 10:57:51 crc kubenswrapper[4678]: I1206 10:57:51.570065 4678 generic.go:334] "Generic (PLEG): container finished" podID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerID="52a269ade91dc4ed8a5efc1327192ce40516eddd39b6add978fd52e91c077477" exitCode=0 Dec 06 10:57:51 crc kubenswrapper[4678]: I1206 10:57:51.569882 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerDied","Data":"6cf498d456910aa38c6f4503596ca4195471c3ab9bc0ea3dd2f50620380ec674"} Dec 06 10:57:51 crc kubenswrapper[4678]: I1206 10:57:51.570099 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerDied","Data":"3164efb7ea39f4016de8c9b1404e49089e5677c33c7a668ebf9bf8e3e1e34d27"} Dec 06 10:57:51 crc kubenswrapper[4678]: I1206 10:57:51.570112 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerDied","Data":"52a269ade91dc4ed8a5efc1327192ce40516eddd39b6add978fd52e91c077477"} Dec 06 10:57:58 crc kubenswrapper[4678]: I1206 10:57:58.645893 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" event={"ID":"4d1feb10-f542-4553-89fc-373d1940ac22","Type":"ContainerStarted","Data":"486a5433f3fcd5aa9866fb1bba4921184592cda13ee965cc3dbd945e0290f2a2"} Dec 06 10:57:58 crc kubenswrapper[4678]: I1206 10:57:58.670368 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" podStartSLOduration=2.6127140129999997 podStartE2EDuration="31.670352077s" podCreationTimestamp="2025-12-06 10:57:27 +0000 UTC" firstStartedPulling="2025-12-06 10:57:28.858690348 +0000 UTC m=+1253.702121787" lastFinishedPulling="2025-12-06 10:57:57.916328412 +0000 UTC m=+1282.759759851" observedRunningTime="2025-12-06 10:57:58.666721232 +0000 UTC m=+1283.510152701" watchObservedRunningTime="2025-12-06 10:57:58.670352077 +0000 UTC m=+1283.513783516" Dec 06 10:58:01 crc kubenswrapper[4678]: I1206 10:58:01.230753 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:58:01 crc kubenswrapper[4678]: I1206 10:58:01.486655 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:58:02 crc kubenswrapper[4678]: I1206 10:58:02.684837 4678 generic.go:334] "Generic (PLEG): container finished" podID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerID="d28c28febe24024c8c3c243eddcc1589c9b2796739ed43f5f21f9b958f0647ed" exitCode=0 Dec 06 10:58:02 crc kubenswrapper[4678]: I1206 10:58:02.685024 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerDied","Data":"d28c28febe24024c8c3c243eddcc1589c9b2796739ed43f5f21f9b958f0647ed"} Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.005857 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193310 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-run-httpd\") pod \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193385 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfspt\" (UniqueName: \"kubernetes.io/projected/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-kube-api-access-hfspt\") pod \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193477 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-sg-core-conf-yaml\") pod \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193637 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-log-httpd\") pod \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193669 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-config-data\") pod \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193715 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-combined-ca-bundle\") pod \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193721 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dd9e74e6-f255-41d1-8ab1-817c83d4fd45" (UID: "dd9e74e6-f255-41d1-8ab1-817c83d4fd45"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.193742 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-scripts\") pod \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\" (UID: \"dd9e74e6-f255-41d1-8ab1-817c83d4fd45\") " Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.195134 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.195349 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dd9e74e6-f255-41d1-8ab1-817c83d4fd45" (UID: "dd9e74e6-f255-41d1-8ab1-817c83d4fd45"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.205680 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-kube-api-access-hfspt" (OuterVolumeSpecName: "kube-api-access-hfspt") pod "dd9e74e6-f255-41d1-8ab1-817c83d4fd45" (UID: "dd9e74e6-f255-41d1-8ab1-817c83d4fd45"). InnerVolumeSpecName "kube-api-access-hfspt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.215730 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-scripts" (OuterVolumeSpecName: "scripts") pod "dd9e74e6-f255-41d1-8ab1-817c83d4fd45" (UID: "dd9e74e6-f255-41d1-8ab1-817c83d4fd45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.262645 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dd9e74e6-f255-41d1-8ab1-817c83d4fd45" (UID: "dd9e74e6-f255-41d1-8ab1-817c83d4fd45"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.297041 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.297074 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.297083 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfspt\" (UniqueName: \"kubernetes.io/projected/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-kube-api-access-hfspt\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.297095 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.331675 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dd9e74e6-f255-41d1-8ab1-817c83d4fd45" (UID: "dd9e74e6-f255-41d1-8ab1-817c83d4fd45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.337462 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-config-data" (OuterVolumeSpecName: "config-data") pod "dd9e74e6-f255-41d1-8ab1-817c83d4fd45" (UID: "dd9e74e6-f255-41d1-8ab1-817c83d4fd45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.398263 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.398296 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd9e74e6-f255-41d1-8ab1-817c83d4fd45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.420449 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.511020 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d77bcbbf8-cfc4l" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.594721 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b97647d4d-lzbd6"] Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.696285 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon-log" containerID="cri-o://d4163c241eb326f5ee5663d607080cbbfd1967af2e9381c39c27f1ceaeebb861" gracePeriod=30 Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.696963 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.699108 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" containerID="cri-o://c7685a17cd3751529dee7b1765aae00ffa7c22bf0283855f7f4a93abf135d645" gracePeriod=30 Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.699268 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dd9e74e6-f255-41d1-8ab1-817c83d4fd45","Type":"ContainerDied","Data":"ece27320d31f716a330bcbef5153eacd6982fda566e8c0933702883a94ededeb"} Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.699306 4678 scope.go:117] "RemoveContainer" containerID="6cf498d456910aa38c6f4503596ca4195471c3ab9bc0ea3dd2f50620380ec674" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.734547 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.751270 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.758781 4678 scope.go:117] "RemoveContainer" containerID="3164efb7ea39f4016de8c9b1404e49089e5677c33c7a668ebf9bf8e3e1e34d27" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787062 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:03 crc kubenswrapper[4678]: E1206 10:58:03.787560 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="sg-core" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787586 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="sg-core" Dec 06 10:58:03 crc kubenswrapper[4678]: E1206 10:58:03.787614 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-notification-agent" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787622 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-notification-agent" Dec 06 10:58:03 crc kubenswrapper[4678]: E1206 10:58:03.787644 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="proxy-httpd" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787653 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="proxy-httpd" Dec 06 10:58:03 crc kubenswrapper[4678]: E1206 10:58:03.787686 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-central-agent" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787694 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-central-agent" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787895 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-notification-agent" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787921 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="ceilometer-central-agent" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787931 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="sg-core" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.787952 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" containerName="proxy-httpd" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.789651 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.798862 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.807507 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.809476 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.830537 4678 scope.go:117] "RemoveContainer" containerID="52a269ade91dc4ed8a5efc1327192ce40516eddd39b6add978fd52e91c077477" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.871182 4678 scope.go:117] "RemoveContainer" containerID="d28c28febe24024c8c3c243eddcc1589c9b2796739ed43f5f21f9b958f0647ed" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.915337 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.915384 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-log-httpd\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.915408 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-run-httpd\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.915432 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28rr2\" (UniqueName: \"kubernetes.io/projected/1c900c1e-448b-4e4f-abec-a934c6387745-kube-api-access-28rr2\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.915461 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-config-data\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.916170 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:03 crc kubenswrapper[4678]: I1206 10:58:03.916241 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-scripts\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.017881 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.017998 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-scripts\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.018040 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.018062 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-log-httpd\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.018085 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-run-httpd\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.018116 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28rr2\" (UniqueName: \"kubernetes.io/projected/1c900c1e-448b-4e4f-abec-a934c6387745-kube-api-access-28rr2\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.018146 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-config-data\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.020721 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-run-httpd\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.021026 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-log-httpd\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.024127 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.025208 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-config-data\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.025253 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.026237 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-scripts\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.043827 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28rr2\" (UniqueName: \"kubernetes.io/projected/1c900c1e-448b-4e4f-abec-a934c6387745-kube-api-access-28rr2\") pod \"ceilometer-0\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.129977 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:04 crc kubenswrapper[4678]: W1206 10:58:04.701432 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c900c1e_448b_4e4f_abec_a934c6387745.slice/crio-4d4958a21c78a6c3138ddea5a430e2f4221933764cf0814b66f169368fa23c69 WatchSource:0}: Error finding container 4d4958a21c78a6c3138ddea5a430e2f4221933764cf0814b66f169368fa23c69: Status 404 returned error can't find the container with id 4d4958a21c78a6c3138ddea5a430e2f4221933764cf0814b66f169368fa23c69 Dec 06 10:58:04 crc kubenswrapper[4678]: I1206 10:58:04.706317 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:05 crc kubenswrapper[4678]: I1206 10:58:05.487255 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd9e74e6-f255-41d1-8ab1-817c83d4fd45" path="/var/lib/kubelet/pods/dd9e74e6-f255-41d1-8ab1-817c83d4fd45/volumes" Dec 06 10:58:05 crc kubenswrapper[4678]: I1206 10:58:05.726408 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerStarted","Data":"39f21ec98e9fe3ae2eab904f8be184a9564a0a314cab859458f98488baaa546c"} Dec 06 10:58:05 crc kubenswrapper[4678]: I1206 10:58:05.726753 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerStarted","Data":"4d4958a21c78a6c3138ddea5a430e2f4221933764cf0814b66f169368fa23c69"} Dec 06 10:58:06 crc kubenswrapper[4678]: I1206 10:58:06.740954 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerStarted","Data":"bf57c41d9ca016dcaabdcc5c1cd2c2cb475ef019c6808a1d5e7922a14c7e17c1"} Dec 06 10:58:06 crc kubenswrapper[4678]: I1206 10:58:06.741251 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerStarted","Data":"a87ea3eb2113bbda1f5cfac425fb57b3cabc6c6a501235c03e2d09c5884b2983"} Dec 06 10:58:07 crc kubenswrapper[4678]: I1206 10:58:07.753135 4678 generic.go:334] "Generic (PLEG): container finished" podID="c3692143-6629-40ba-984b-187126e4a2ed" containerID="c7685a17cd3751529dee7b1765aae00ffa7c22bf0283855f7f4a93abf135d645" exitCode=0 Dec 06 10:58:07 crc kubenswrapper[4678]: I1206 10:58:07.753624 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerDied","Data":"c7685a17cd3751529dee7b1765aae00ffa7c22bf0283855f7f4a93abf135d645"} Dec 06 10:58:07 crc kubenswrapper[4678]: I1206 10:58:07.753670 4678 scope.go:117] "RemoveContainer" containerID="389817da78af88665e5f2000c353529adec4c3df99e02f2713ce867a856272b2" Dec 06 10:58:08 crc kubenswrapper[4678]: I1206 10:58:08.791675 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1cb14647-bb80-422d-8f59-985a70bf14c3" containerName="galera" probeResult="failure" output="command timed out" Dec 06 10:58:08 crc kubenswrapper[4678]: I1206 10:58:08.802630 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1cb14647-bb80-422d-8f59-985a70bf14c3" containerName="galera" probeResult="failure" output="command timed out" Dec 06 10:58:08 crc kubenswrapper[4678]: I1206 10:58:08.977398 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:58:09 crc kubenswrapper[4678]: I1206 10:58:09.831870 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerStarted","Data":"cc7844f589adf8baeaac0a953c9c63afb6b341d6f73a50cefd4efa2d189be859"} Dec 06 10:58:09 crc kubenswrapper[4678]: I1206 10:58:09.832240 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 10:58:09 crc kubenswrapper[4678]: I1206 10:58:09.854793 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.526967073 podStartE2EDuration="6.854770833s" podCreationTimestamp="2025-12-06 10:58:03 +0000 UTC" firstStartedPulling="2025-12-06 10:58:04.704421357 +0000 UTC m=+1289.547852796" lastFinishedPulling="2025-12-06 10:58:09.032225117 +0000 UTC m=+1293.875656556" observedRunningTime="2025-12-06 10:58:09.849169576 +0000 UTC m=+1294.692601015" watchObservedRunningTime="2025-12-06 10:58:09.854770833 +0000 UTC m=+1294.698202272" Dec 06 10:58:12 crc kubenswrapper[4678]: I1206 10:58:12.861643 4678 generic.go:334] "Generic (PLEG): container finished" podID="4d1feb10-f542-4553-89fc-373d1940ac22" containerID="486a5433f3fcd5aa9866fb1bba4921184592cda13ee965cc3dbd945e0290f2a2" exitCode=0 Dec 06 10:58:12 crc kubenswrapper[4678]: I1206 10:58:12.861679 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" event={"ID":"4d1feb10-f542-4553-89fc-373d1940ac22","Type":"ContainerDied","Data":"486a5433f3fcd5aa9866fb1bba4921184592cda13ee965cc3dbd945e0290f2a2"} Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.307071 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.375959 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-combined-ca-bundle\") pod \"4d1feb10-f542-4553-89fc-373d1940ac22\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.376043 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxz48\" (UniqueName: \"kubernetes.io/projected/4d1feb10-f542-4553-89fc-373d1940ac22-kube-api-access-sxz48\") pod \"4d1feb10-f542-4553-89fc-373d1940ac22\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.376220 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-config-data\") pod \"4d1feb10-f542-4553-89fc-373d1940ac22\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.376252 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-scripts\") pod \"4d1feb10-f542-4553-89fc-373d1940ac22\" (UID: \"4d1feb10-f542-4553-89fc-373d1940ac22\") " Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.384272 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d1feb10-f542-4553-89fc-373d1940ac22-kube-api-access-sxz48" (OuterVolumeSpecName: "kube-api-access-sxz48") pod "4d1feb10-f542-4553-89fc-373d1940ac22" (UID: "4d1feb10-f542-4553-89fc-373d1940ac22"). InnerVolumeSpecName "kube-api-access-sxz48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.388016 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-scripts" (OuterVolumeSpecName: "scripts") pod "4d1feb10-f542-4553-89fc-373d1940ac22" (UID: "4d1feb10-f542-4553-89fc-373d1940ac22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.411344 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-config-data" (OuterVolumeSpecName: "config-data") pod "4d1feb10-f542-4553-89fc-373d1940ac22" (UID: "4d1feb10-f542-4553-89fc-373d1940ac22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.413690 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d1feb10-f542-4553-89fc-373d1940ac22" (UID: "4d1feb10-f542-4553-89fc-373d1940ac22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.477781 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.477814 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxz48\" (UniqueName: \"kubernetes.io/projected/4d1feb10-f542-4553-89fc-373d1940ac22-kube-api-access-sxz48\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.477830 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.477840 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d1feb10-f542-4553-89fc-373d1940ac22-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.887942 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" event={"ID":"4d1feb10-f542-4553-89fc-373d1940ac22","Type":"ContainerDied","Data":"787bd25dc6d4f8709128e940721c0fa0ba0a5f5b1811f75b6325a9a6309b4ccb"} Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.887985 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xdm9z" Dec 06 10:58:14 crc kubenswrapper[4678]: I1206 10:58:14.888004 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="787bd25dc6d4f8709128e940721c0fa0ba0a5f5b1811f75b6325a9a6309b4ccb" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.006996 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 10:58:15 crc kubenswrapper[4678]: E1206 10:58:15.007444 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d1feb10-f542-4553-89fc-373d1940ac22" containerName="nova-cell0-conductor-db-sync" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.007466 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d1feb10-f542-4553-89fc-373d1940ac22" containerName="nova-cell0-conductor-db-sync" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.007751 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d1feb10-f542-4553-89fc-373d1940ac22" containerName="nova-cell0-conductor-db-sync" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.008532 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.011824 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-t9whj" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.014188 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.034786 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.088050 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92133d6-4bca-447f-9b4c-a1e776a4cb40-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.088221 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92133d6-4bca-447f-9b4c-a1e776a4cb40-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.088342 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26z68\" (UniqueName: \"kubernetes.io/projected/a92133d6-4bca-447f-9b4c-a1e776a4cb40-kube-api-access-26z68\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.189455 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92133d6-4bca-447f-9b4c-a1e776a4cb40-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.189547 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26z68\" (UniqueName: \"kubernetes.io/projected/a92133d6-4bca-447f-9b4c-a1e776a4cb40-kube-api-access-26z68\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.189666 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92133d6-4bca-447f-9b4c-a1e776a4cb40-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.194766 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a92133d6-4bca-447f-9b4c-a1e776a4cb40-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.194995 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a92133d6-4bca-447f-9b4c-a1e776a4cb40-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.207355 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26z68\" (UniqueName: \"kubernetes.io/projected/a92133d6-4bca-447f-9b4c-a1e776a4cb40-kube-api-access-26z68\") pod \"nova-cell0-conductor-0\" (UID: \"a92133d6-4bca-447f-9b4c-a1e776a4cb40\") " pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.327718 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.757076 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 06 10:58:15 crc kubenswrapper[4678]: W1206 10:58:15.761943 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda92133d6_4bca_447f_9b4c_a1e776a4cb40.slice/crio-42efca93ba2057b7b9d63e3f602208aa622132b142282e402598fd32dea8607f WatchSource:0}: Error finding container 42efca93ba2057b7b9d63e3f602208aa622132b142282e402598fd32dea8607f: Status 404 returned error can't find the container with id 42efca93ba2057b7b9d63e3f602208aa622132b142282e402598fd32dea8607f Dec 06 10:58:15 crc kubenswrapper[4678]: I1206 10:58:15.897651 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a92133d6-4bca-447f-9b4c-a1e776a4cb40","Type":"ContainerStarted","Data":"42efca93ba2057b7b9d63e3f602208aa622132b142282e402598fd32dea8607f"} Dec 06 10:58:16 crc kubenswrapper[4678]: I1206 10:58:16.910665 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"a92133d6-4bca-447f-9b4c-a1e776a4cb40","Type":"ContainerStarted","Data":"3d3d80517499c08fe34c0c0ac4061359e5d7523739b345d200f3a472b8176084"} Dec 06 10:58:16 crc kubenswrapper[4678]: I1206 10:58:16.911284 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:16 crc kubenswrapper[4678]: I1206 10:58:16.932680 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.932655235 podStartE2EDuration="2.932655235s" podCreationTimestamp="2025-12-06 10:58:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:16.93024688 +0000 UTC m=+1301.773678329" watchObservedRunningTime="2025-12-06 10:58:16.932655235 +0000 UTC m=+1301.776086674" Dec 06 10:58:18 crc kubenswrapper[4678]: I1206 10:58:18.977637 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.359221 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.827770 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-thzqq"] Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.829542 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.834201 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.834480 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.846629 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-thzqq"] Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.903560 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.903615 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2jnz\" (UniqueName: \"kubernetes.io/projected/dea23b4d-e666-4ba6-954d-c07956044128-kube-api-access-h2jnz\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.903639 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-scripts\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:20 crc kubenswrapper[4678]: I1206 10:58:20.903794 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-config-data\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.005920 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-config-data\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.006044 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.006073 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2jnz\" (UniqueName: \"kubernetes.io/projected/dea23b4d-e666-4ba6-954d-c07956044128-kube-api-access-h2jnz\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.006095 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-scripts\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.023560 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.023725 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-scripts\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.023651 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-config-data\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.026956 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.028389 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.034122 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.062092 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.123080 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.123176 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-config-data\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.123196 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qblxl\" (UniqueName: \"kubernetes.io/projected/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-kube-api-access-qblxl\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.127062 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2jnz\" (UniqueName: \"kubernetes.io/projected/dea23b4d-e666-4ba6-954d-c07956044128-kube-api-access-h2jnz\") pod \"nova-cell0-cell-mapping-thzqq\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.152142 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.204591 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.222563 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.224237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.224287 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-config-data\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.224307 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qblxl\" (UniqueName: \"kubernetes.io/projected/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-kube-api-access-qblxl\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.234594 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.237619 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-config-data\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.238612 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.301800 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.315207 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qblxl\" (UniqueName: \"kubernetes.io/projected/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-kube-api-access-qblxl\") pod \"nova-scheduler-0\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.332104 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.333327 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.334623 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-config-data\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.334674 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5031a03-e205-42a3-8c23-6911af163988-logs\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.334763 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.334791 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqd9r\" (UniqueName: \"kubernetes.io/projected/e5031a03-e205-42a3-8c23-6911af163988-kube-api-access-hqd9r\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.341140 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.389559 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.391171 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.397887 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.423729 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.435905 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.437733 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-config-data\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.437770 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5031a03-e205-42a3-8c23-6911af163988-logs\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.437822 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.437843 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpz6k\" (UniqueName: \"kubernetes.io/projected/5a9b1e8b-82ef-4abb-aff7-751e89de309f-kube-api-access-qpz6k\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.437873 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.437896 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqd9r\" (UniqueName: \"kubernetes.io/projected/e5031a03-e205-42a3-8c23-6911af163988-kube-api-access-hqd9r\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.437932 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.443371 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5031a03-e205-42a3-8c23-6911af163988-logs\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.446130 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-config-data\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.456607 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.486439 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqd9r\" (UniqueName: \"kubernetes.io/projected/e5031a03-e205-42a3-8c23-6911af163988-kube-api-access-hqd9r\") pod \"nova-api-0\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.490245 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.515237 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-9brh9"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.518674 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.541886 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-logs\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.542031 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.542716 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84hsr\" (UniqueName: \"kubernetes.io/projected/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-kube-api-access-84hsr\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.542859 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.542943 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpz6k\" (UniqueName: \"kubernetes.io/projected/5a9b1e8b-82ef-4abb-aff7-751e89de309f-kube-api-access-qpz6k\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.543113 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.543226 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-config-data\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.555403 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.556955 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.565040 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpz6k\" (UniqueName: \"kubernetes.io/projected/5a9b1e8b-82ef-4abb-aff7-751e89de309f-kube-api-access-qpz6k\") pod \"nova-cell1-novncproxy-0\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.581622 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-9brh9"] Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.628897 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.644764 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.644817 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-config-data\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.644848 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-logs\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.644919 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-svc\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.644946 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84hsr\" (UniqueName: \"kubernetes.io/projected/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-kube-api-access-84hsr\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.644968 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-config\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.644985 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9sxx\" (UniqueName: \"kubernetes.io/projected/69716076-6f0e-4432-91fc-82ae535dd96b-kube-api-access-d9sxx\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.645012 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.645054 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.645082 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.647222 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-logs\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.649952 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.653314 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-config-data\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.666095 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84hsr\" (UniqueName: \"kubernetes.io/projected/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-kube-api-access-84hsr\") pod \"nova-metadata-0\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.689034 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.726943 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.747535 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-svc\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.747606 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9sxx\" (UniqueName: \"kubernetes.io/projected/69716076-6f0e-4432-91fc-82ae535dd96b-kube-api-access-d9sxx\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.747627 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-config\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.747648 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.747692 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.747722 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.748964 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.749026 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-svc\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.749647 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.750172 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.751451 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-config\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.775716 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9sxx\" (UniqueName: \"kubernetes.io/projected/69716076-6f0e-4432-91fc-82ae535dd96b-kube-api-access-d9sxx\") pod \"dnsmasq-dns-bccf8f775-9brh9\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.861039 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:21 crc kubenswrapper[4678]: I1206 10:58:21.978030 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-thzqq"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.143087 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.177366 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.390057 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.413991 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.526251 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-df7h2"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.527885 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.530670 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.534833 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.547634 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-df7h2"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.564552 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.655581 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-9brh9"] Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.683803 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8jhj\" (UniqueName: \"kubernetes.io/projected/02942adf-5e19-482e-a76a-16d40ec52416-kube-api-access-s8jhj\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.684077 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.684158 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-scripts\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.684188 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-config-data\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.785673 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-config-data\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.785800 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8jhj\" (UniqueName: \"kubernetes.io/projected/02942adf-5e19-482e-a76a-16d40ec52416-kube-api-access-s8jhj\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.786168 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.786720 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-scripts\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.801168 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-scripts\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.801298 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.801850 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-config-data\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.810760 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8jhj\" (UniqueName: \"kubernetes.io/projected/02942adf-5e19-482e-a76a-16d40ec52416-kube-api-access-s8jhj\") pod \"nova-cell1-conductor-db-sync-df7h2\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.854080 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.986640 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" event={"ID":"69716076-6f0e-4432-91fc-82ae535dd96b","Type":"ContainerDied","Data":"a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe"} Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.987424 4678 generic.go:334] "Generic (PLEG): container finished" podID="69716076-6f0e-4432-91fc-82ae535dd96b" containerID="a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe" exitCode=0 Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.987530 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" event={"ID":"69716076-6f0e-4432-91fc-82ae535dd96b","Type":"ContainerStarted","Data":"fbc11a2bb670858f8c7ab950319dc9f1097fd08a25221c6dda10006aec4747a7"} Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.990074 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be17b2be-bbb5-49c1-8760-6ac7a76a70f1","Type":"ContainerStarted","Data":"0d825bb9ed61984e4fffc7da1de77de8517ce42c659e81d075927ab2abdca591"} Dec 06 10:58:22 crc kubenswrapper[4678]: I1206 10:58:22.995944 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5031a03-e205-42a3-8c23-6911af163988","Type":"ContainerStarted","Data":"ece93022be59381794959560fa685f54a6afd1fb3907943e7b9ed077be66b6a2"} Dec 06 10:58:23 crc kubenswrapper[4678]: I1206 10:58:23.014231 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c7ff6af4-3ad5-498d-8d54-39105ba9f01d","Type":"ContainerStarted","Data":"5142ba13cccf95873ebac3a9cea310d0a922c442c39bbbb283bdd93d832ce0bf"} Dec 06 10:58:23 crc kubenswrapper[4678]: I1206 10:58:23.036335 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5a9b1e8b-82ef-4abb-aff7-751e89de309f","Type":"ContainerStarted","Data":"0d44d8ad2b506e6e56027ad477a79c74a9b78515d303469fa89346264e3888dd"} Dec 06 10:58:23 crc kubenswrapper[4678]: I1206 10:58:23.042137 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-thzqq" event={"ID":"dea23b4d-e666-4ba6-954d-c07956044128","Type":"ContainerStarted","Data":"25ffb982290c4206502d8c86b1733a1e6c83d918b7c462e18e09ff55e90541f3"} Dec 06 10:58:23 crc kubenswrapper[4678]: I1206 10:58:23.042182 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-thzqq" event={"ID":"dea23b4d-e666-4ba6-954d-c07956044128","Type":"ContainerStarted","Data":"eec566da813102137aa2e71afe864e3d61ddba7a6731aed8ec151caa330e055a"} Dec 06 10:58:23 crc kubenswrapper[4678]: I1206 10:58:23.105408 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-thzqq" podStartSLOduration=3.105387247 podStartE2EDuration="3.105387247s" podCreationTimestamp="2025-12-06 10:58:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:23.06930926 +0000 UTC m=+1307.912740699" watchObservedRunningTime="2025-12-06 10:58:23.105387247 +0000 UTC m=+1307.948818686" Dec 06 10:58:23 crc kubenswrapper[4678]: I1206 10:58:23.287925 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-df7h2"] Dec 06 10:58:24 crc kubenswrapper[4678]: I1206 10:58:24.055108 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-df7h2" event={"ID":"02942adf-5e19-482e-a76a-16d40ec52416","Type":"ContainerStarted","Data":"0b127b3a998bc68c192239209fdbf6fee3765fd1f2d5bb7a1fd29011bbf759b7"} Dec 06 10:58:24 crc kubenswrapper[4678]: I1206 10:58:24.055377 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-df7h2" event={"ID":"02942adf-5e19-482e-a76a-16d40ec52416","Type":"ContainerStarted","Data":"c47f47fa8fd60ac74b624fbee274c756e3adff47c8f2f327b22a80ae4c5b4b76"} Dec 06 10:58:24 crc kubenswrapper[4678]: I1206 10:58:24.064893 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" event={"ID":"69716076-6f0e-4432-91fc-82ae535dd96b","Type":"ContainerStarted","Data":"e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12"} Dec 06 10:58:24 crc kubenswrapper[4678]: I1206 10:58:24.064980 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:24 crc kubenswrapper[4678]: I1206 10:58:24.089929 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-df7h2" podStartSLOduration=2.08990753 podStartE2EDuration="2.08990753s" podCreationTimestamp="2025-12-06 10:58:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:24.084984415 +0000 UTC m=+1308.928415854" watchObservedRunningTime="2025-12-06 10:58:24.08990753 +0000 UTC m=+1308.933338969" Dec 06 10:58:24 crc kubenswrapper[4678]: I1206 10:58:24.110173 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" podStartSLOduration=3.110157868 podStartE2EDuration="3.110157868s" podCreationTimestamp="2025-12-06 10:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:24.107796654 +0000 UTC m=+1308.951228093" watchObservedRunningTime="2025-12-06 10:58:24.110157868 +0000 UTC m=+1308.953589307" Dec 06 10:58:25 crc kubenswrapper[4678]: I1206 10:58:25.207691 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:25 crc kubenswrapper[4678]: I1206 10:58:25.232141 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:27 crc kubenswrapper[4678]: I1206 10:58:27.111677 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c7ff6af4-3ad5-498d-8d54-39105ba9f01d","Type":"ContainerStarted","Data":"15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69"} Dec 06 10:58:27 crc kubenswrapper[4678]: I1206 10:58:27.115697 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5a9b1e8b-82ef-4abb-aff7-751e89de309f","Type":"ContainerStarted","Data":"3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2"} Dec 06 10:58:27 crc kubenswrapper[4678]: I1206 10:58:27.115813 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="5a9b1e8b-82ef-4abb-aff7-751e89de309f" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2" gracePeriod=30 Dec 06 10:58:27 crc kubenswrapper[4678]: I1206 10:58:27.117705 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be17b2be-bbb5-49c1-8760-6ac7a76a70f1","Type":"ContainerStarted","Data":"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a"} Dec 06 10:58:27 crc kubenswrapper[4678]: I1206 10:58:27.125071 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5031a03-e205-42a3-8c23-6911af163988","Type":"ContainerStarted","Data":"9f1c3e4b8b048fa251aca2d0e8c0d3446318619e9e59455068c5eafc26e29dda"} Dec 06 10:58:27 crc kubenswrapper[4678]: I1206 10:58:27.142800 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.658384852 podStartE2EDuration="7.14278053s" podCreationTimestamp="2025-12-06 10:58:20 +0000 UTC" firstStartedPulling="2025-12-06 10:58:22.177166749 +0000 UTC m=+1307.020598188" lastFinishedPulling="2025-12-06 10:58:26.661562427 +0000 UTC m=+1311.504993866" observedRunningTime="2025-12-06 10:58:27.138598078 +0000 UTC m=+1311.982029517" watchObservedRunningTime="2025-12-06 10:58:27.14278053 +0000 UTC m=+1311.986211969" Dec 06 10:58:27 crc kubenswrapper[4678]: I1206 10:58:27.161319 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.933333782 podStartE2EDuration="6.161299664s" podCreationTimestamp="2025-12-06 10:58:21 +0000 UTC" firstStartedPulling="2025-12-06 10:58:22.433700038 +0000 UTC m=+1307.277131467" lastFinishedPulling="2025-12-06 10:58:26.66166591 +0000 UTC m=+1311.505097349" observedRunningTime="2025-12-06 10:58:27.15356764 +0000 UTC m=+1311.996999079" watchObservedRunningTime="2025-12-06 10:58:27.161299664 +0000 UTC m=+1312.004731103" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.136182 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5031a03-e205-42a3-8c23-6911af163988","Type":"ContainerStarted","Data":"e32906004e2a26992632358a2bd005f0d28afb3ddec91dc760ea47752c1825a8"} Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.140866 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-log" containerID="cri-o://db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a" gracePeriod=30 Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.140983 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-metadata" containerID="cri-o://079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b" gracePeriod=30 Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.141181 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be17b2be-bbb5-49c1-8760-6ac7a76a70f1","Type":"ContainerStarted","Data":"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b"} Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.155786 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.889099128 podStartE2EDuration="7.15577139s" podCreationTimestamp="2025-12-06 10:58:21 +0000 UTC" firstStartedPulling="2025-12-06 10:58:22.406648635 +0000 UTC m=+1307.250080064" lastFinishedPulling="2025-12-06 10:58:26.673320887 +0000 UTC m=+1311.516752326" observedRunningTime="2025-12-06 10:58:28.150701751 +0000 UTC m=+1312.994133190" watchObservedRunningTime="2025-12-06 10:58:28.15577139 +0000 UTC m=+1312.999202829" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.205612 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.116760857 podStartE2EDuration="7.205593281s" podCreationTimestamp="2025-12-06 10:58:21 +0000 UTC" firstStartedPulling="2025-12-06 10:58:22.574275391 +0000 UTC m=+1307.417706830" lastFinishedPulling="2025-12-06 10:58:26.663107815 +0000 UTC m=+1311.506539254" observedRunningTime="2025-12-06 10:58:28.198424466 +0000 UTC m=+1313.041855915" watchObservedRunningTime="2025-12-06 10:58:28.205593281 +0000 UTC m=+1313.049024720" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.749680 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.818439 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-combined-ca-bundle\") pod \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.818542 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-config-data\") pod \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.818701 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-logs\") pod \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.818795 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84hsr\" (UniqueName: \"kubernetes.io/projected/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-kube-api-access-84hsr\") pod \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\" (UID: \"be17b2be-bbb5-49c1-8760-6ac7a76a70f1\") " Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.820260 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-logs" (OuterVolumeSpecName: "logs") pod "be17b2be-bbb5-49c1-8760-6ac7a76a70f1" (UID: "be17b2be-bbb5-49c1-8760-6ac7a76a70f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.846709 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-kube-api-access-84hsr" (OuterVolumeSpecName: "kube-api-access-84hsr") pod "be17b2be-bbb5-49c1-8760-6ac7a76a70f1" (UID: "be17b2be-bbb5-49c1-8760-6ac7a76a70f1"). InnerVolumeSpecName "kube-api-access-84hsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.881103 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "be17b2be-bbb5-49c1-8760-6ac7a76a70f1" (UID: "be17b2be-bbb5-49c1-8760-6ac7a76a70f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.894044 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-config-data" (OuterVolumeSpecName: "config-data") pod "be17b2be-bbb5-49c1-8760-6ac7a76a70f1" (UID: "be17b2be-bbb5-49c1-8760-6ac7a76a70f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.922433 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.922498 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84hsr\" (UniqueName: \"kubernetes.io/projected/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-kube-api-access-84hsr\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.922514 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.922524 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be17b2be-bbb5-49c1-8760-6ac7a76a70f1-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.977249 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-5b97647d4d-lzbd6" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 06 10:58:28 crc kubenswrapper[4678]: I1206 10:58:28.977376 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.150089 4678 generic.go:334] "Generic (PLEG): container finished" podID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerID="079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b" exitCode=0 Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.150117 4678 generic.go:334] "Generic (PLEG): container finished" podID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerID="db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a" exitCode=143 Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.150161 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.150174 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be17b2be-bbb5-49c1-8760-6ac7a76a70f1","Type":"ContainerDied","Data":"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b"} Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.150210 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be17b2be-bbb5-49c1-8760-6ac7a76a70f1","Type":"ContainerDied","Data":"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a"} Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.150219 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be17b2be-bbb5-49c1-8760-6ac7a76a70f1","Type":"ContainerDied","Data":"0d825bb9ed61984e4fffc7da1de77de8517ce42c659e81d075927ab2abdca591"} Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.150235 4678 scope.go:117] "RemoveContainer" containerID="079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.180410 4678 scope.go:117] "RemoveContainer" containerID="db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.185766 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.198229 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.205833 4678 scope.go:117] "RemoveContainer" containerID="079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.208301 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:29 crc kubenswrapper[4678]: E1206 10:58:29.208694 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-log" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.208710 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-log" Dec 06 10:58:29 crc kubenswrapper[4678]: E1206 10:58:29.208726 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-metadata" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.208732 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-metadata" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.208899 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-log" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.208923 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" containerName="nova-metadata-metadata" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.209883 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: E1206 10:58:29.210980 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b\": container with ID starting with 079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b not found: ID does not exist" containerID="079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.211008 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b"} err="failed to get container status \"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b\": rpc error: code = NotFound desc = could not find container \"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b\": container with ID starting with 079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b not found: ID does not exist" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.211032 4678 scope.go:117] "RemoveContainer" containerID="db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a" Dec 06 10:58:29 crc kubenswrapper[4678]: E1206 10:58:29.211983 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a\": container with ID starting with db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a not found: ID does not exist" containerID="db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.212024 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a"} err="failed to get container status \"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a\": rpc error: code = NotFound desc = could not find container \"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a\": container with ID starting with db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a not found: ID does not exist" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.212052 4678 scope.go:117] "RemoveContainer" containerID="079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.212126 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.212232 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.212335 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b"} err="failed to get container status \"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b\": rpc error: code = NotFound desc = could not find container \"079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b\": container with ID starting with 079a3816307450dbea102b387292ace7b5d2e1e4749ff7d9a9624875c5ffcd1b not found: ID does not exist" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.212353 4678 scope.go:117] "RemoveContainer" containerID="db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.213937 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a"} err="failed to get container status \"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a\": rpc error: code = NotFound desc = could not find container \"db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a\": container with ID starting with db004ebce2d4d11ffac0037caf442a25705cc90820044b363cadba8cc2ada61a not found: ID does not exist" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.227601 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.329964 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.329997 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-config-data\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.330068 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/298ed19a-5247-4a42-ae20-a00a511db3ba-logs\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.330092 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqlbk\" (UniqueName: \"kubernetes.io/projected/298ed19a-5247-4a42-ae20-a00a511db3ba-kube-api-access-zqlbk\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.330118 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.431512 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.431566 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-config-data\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.431638 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/298ed19a-5247-4a42-ae20-a00a511db3ba-logs\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.431665 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqlbk\" (UniqueName: \"kubernetes.io/projected/298ed19a-5247-4a42-ae20-a00a511db3ba-kube-api-access-zqlbk\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.431689 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.432330 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/298ed19a-5247-4a42-ae20-a00a511db3ba-logs\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.437948 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.438437 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.442442 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-config-data\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.451757 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqlbk\" (UniqueName: \"kubernetes.io/projected/298ed19a-5247-4a42-ae20-a00a511db3ba-kube-api-access-zqlbk\") pod \"nova-metadata-0\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " pod="openstack/nova-metadata-0" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.486466 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be17b2be-bbb5-49c1-8760-6ac7a76a70f1" path="/var/lib/kubelet/pods/be17b2be-bbb5-49c1-8760-6ac7a76a70f1/volumes" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.505191 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.505248 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:58:29 crc kubenswrapper[4678]: I1206 10:58:29.532111 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:30 crc kubenswrapper[4678]: I1206 10:58:30.035436 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:30 crc kubenswrapper[4678]: I1206 10:58:30.161907 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"298ed19a-5247-4a42-ae20-a00a511db3ba","Type":"ContainerStarted","Data":"de96ff89bb591282200b69e221b91fa59912be1306b4c9b23e5f19109b4caa33"} Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.185296 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"298ed19a-5247-4a42-ae20-a00a511db3ba","Type":"ContainerStarted","Data":"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85"} Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.186579 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"298ed19a-5247-4a42-ae20-a00a511db3ba","Type":"ContainerStarted","Data":"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e"} Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.222773 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.222752745 podStartE2EDuration="2.222752745s" podCreationTimestamp="2025-12-06 10:58:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:31.209481967 +0000 UTC m=+1316.052913416" watchObservedRunningTime="2025-12-06 10:58:31.222752745 +0000 UTC m=+1316.066184194" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.491186 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.491250 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.518559 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.629750 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.630156 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.690272 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.863730 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.929821 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-mkvr4"] Dec 06 10:58:31 crc kubenswrapper[4678]: I1206 10:58:31.930058 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" podUID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerName="dnsmasq-dns" containerID="cri-o://d554c6311667f0da84e4fa15eb86691184370fbf4716480b062539f4ef603e82" gracePeriod=10 Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.221268 4678 generic.go:334] "Generic (PLEG): container finished" podID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerID="d554c6311667f0da84e4fa15eb86691184370fbf4716480b062539f4ef603e82" exitCode=0 Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.221330 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" event={"ID":"eb1e07f0-3034-47d2-9143-82058df6d73c","Type":"ContainerDied","Data":"d554c6311667f0da84e4fa15eb86691184370fbf4716480b062539f4ef603e82"} Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.311880 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.712688 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.713158 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.722845 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.808571 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-config\") pod \"eb1e07f0-3034-47d2-9143-82058df6d73c\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.808633 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-svc\") pod \"eb1e07f0-3034-47d2-9143-82058df6d73c\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.808714 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-nb\") pod \"eb1e07f0-3034-47d2-9143-82058df6d73c\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.808765 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-swift-storage-0\") pod \"eb1e07f0-3034-47d2-9143-82058df6d73c\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.808863 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-sb\") pod \"eb1e07f0-3034-47d2-9143-82058df6d73c\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.809053 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn9dh\" (UniqueName: \"kubernetes.io/projected/eb1e07f0-3034-47d2-9143-82058df6d73c-kube-api-access-wn9dh\") pod \"eb1e07f0-3034-47d2-9143-82058df6d73c\" (UID: \"eb1e07f0-3034-47d2-9143-82058df6d73c\") " Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.834194 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1e07f0-3034-47d2-9143-82058df6d73c-kube-api-access-wn9dh" (OuterVolumeSpecName: "kube-api-access-wn9dh") pod "eb1e07f0-3034-47d2-9143-82058df6d73c" (UID: "eb1e07f0-3034-47d2-9143-82058df6d73c"). InnerVolumeSpecName "kube-api-access-wn9dh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.911551 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn9dh\" (UniqueName: \"kubernetes.io/projected/eb1e07f0-3034-47d2-9143-82058df6d73c-kube-api-access-wn9dh\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.917082 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-config" (OuterVolumeSpecName: "config") pod "eb1e07f0-3034-47d2-9143-82058df6d73c" (UID: "eb1e07f0-3034-47d2-9143-82058df6d73c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.924038 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eb1e07f0-3034-47d2-9143-82058df6d73c" (UID: "eb1e07f0-3034-47d2-9143-82058df6d73c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.930177 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eb1e07f0-3034-47d2-9143-82058df6d73c" (UID: "eb1e07f0-3034-47d2-9143-82058df6d73c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.936027 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eb1e07f0-3034-47d2-9143-82058df6d73c" (UID: "eb1e07f0-3034-47d2-9143-82058df6d73c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:58:32 crc kubenswrapper[4678]: I1206 10:58:32.936658 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eb1e07f0-3034-47d2-9143-82058df6d73c" (UID: "eb1e07f0-3034-47d2-9143-82058df6d73c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.013664 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.013708 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.013723 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.013734 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.013743 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eb1e07f0-3034-47d2-9143-82058df6d73c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.232405 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.232399 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-mkvr4" event={"ID":"eb1e07f0-3034-47d2-9143-82058df6d73c","Type":"ContainerDied","Data":"396de4321be4ab465449d73aaf0b8a4bd09893ea04b23095a64ce28dc90a4cf4"} Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.232513 4678 scope.go:117] "RemoveContainer" containerID="d554c6311667f0da84e4fa15eb86691184370fbf4716480b062539f4ef603e82" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.270527 4678 scope.go:117] "RemoveContainer" containerID="299e677dad807c4dd6e0139b8af97619a409d6baa0b114364570545bcd8ed161" Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.274918 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-mkvr4"] Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.286326 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-mkvr4"] Dec 06 10:58:33 crc kubenswrapper[4678]: I1206 10:58:33.486306 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb1e07f0-3034-47d2-9143-82058df6d73c" path="/var/lib/kubelet/pods/eb1e07f0-3034-47d2-9143-82058df6d73c/volumes" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.136698 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.255299 4678 generic.go:334] "Generic (PLEG): container finished" podID="dea23b4d-e666-4ba6-954d-c07956044128" containerID="25ffb982290c4206502d8c86b1733a1e6c83d918b7c462e18e09ff55e90541f3" exitCode=0 Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.255369 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-thzqq" event={"ID":"dea23b4d-e666-4ba6-954d-c07956044128","Type":"ContainerDied","Data":"25ffb982290c4206502d8c86b1733a1e6c83d918b7c462e18e09ff55e90541f3"} Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.257421 4678 generic.go:334] "Generic (PLEG): container finished" podID="c3692143-6629-40ba-984b-187126e4a2ed" containerID="d4163c241eb326f5ee5663d607080cbbfd1967af2e9381c39c27f1ceaeebb861" exitCode=137 Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.257452 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerDied","Data":"d4163c241eb326f5ee5663d607080cbbfd1967af2e9381c39c27f1ceaeebb861"} Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.532214 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.533699 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.889676 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.948522 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzkqw\" (UniqueName: \"kubernetes.io/projected/c3692143-6629-40ba-984b-187126e4a2ed-kube-api-access-gzkqw\") pod \"c3692143-6629-40ba-984b-187126e4a2ed\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.948663 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3692143-6629-40ba-984b-187126e4a2ed-logs\") pod \"c3692143-6629-40ba-984b-187126e4a2ed\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.948735 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-scripts\") pod \"c3692143-6629-40ba-984b-187126e4a2ed\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.948783 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-secret-key\") pod \"c3692143-6629-40ba-984b-187126e4a2ed\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.948843 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-tls-certs\") pod \"c3692143-6629-40ba-984b-187126e4a2ed\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.948878 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-combined-ca-bundle\") pod \"c3692143-6629-40ba-984b-187126e4a2ed\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.948925 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-config-data\") pod \"c3692143-6629-40ba-984b-187126e4a2ed\" (UID: \"c3692143-6629-40ba-984b-187126e4a2ed\") " Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.949207 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3692143-6629-40ba-984b-187126e4a2ed-logs" (OuterVolumeSpecName: "logs") pod "c3692143-6629-40ba-984b-187126e4a2ed" (UID: "c3692143-6629-40ba-984b-187126e4a2ed"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.959729 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3692143-6629-40ba-984b-187126e4a2ed-kube-api-access-gzkqw" (OuterVolumeSpecName: "kube-api-access-gzkqw") pod "c3692143-6629-40ba-984b-187126e4a2ed" (UID: "c3692143-6629-40ba-984b-187126e4a2ed"). InnerVolumeSpecName "kube-api-access-gzkqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.966439 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c3692143-6629-40ba-984b-187126e4a2ed" (UID: "c3692143-6629-40ba-984b-187126e4a2ed"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:34 crc kubenswrapper[4678]: I1206 10:58:34.989047 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-config-data" (OuterVolumeSpecName: "config-data") pod "c3692143-6629-40ba-984b-187126e4a2ed" (UID: "c3692143-6629-40ba-984b-187126e4a2ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.016675 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3692143-6629-40ba-984b-187126e4a2ed" (UID: "c3692143-6629-40ba-984b-187126e4a2ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.031715 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-scripts" (OuterVolumeSpecName: "scripts") pod "c3692143-6629-40ba-984b-187126e4a2ed" (UID: "c3692143-6629-40ba-984b-187126e4a2ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.046027 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "c3692143-6629-40ba-984b-187126e4a2ed" (UID: "c3692143-6629-40ba-984b-187126e4a2ed"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.050668 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.050700 4678 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.050710 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3692143-6629-40ba-984b-187126e4a2ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.050719 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.050728 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzkqw\" (UniqueName: \"kubernetes.io/projected/c3692143-6629-40ba-984b-187126e4a2ed-kube-api-access-gzkqw\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.050740 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c3692143-6629-40ba-984b-187126e4a2ed-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.050749 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3692143-6629-40ba-984b-187126e4a2ed-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.269196 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5b97647d4d-lzbd6" event={"ID":"c3692143-6629-40ba-984b-187126e4a2ed","Type":"ContainerDied","Data":"835d99378e15f7ba1a46a2c045e9db87f8dbdd6f5874ca7cc09d864908cbb271"} Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.269260 4678 scope.go:117] "RemoveContainer" containerID="c7685a17cd3751529dee7b1765aae00ffa7c22bf0283855f7f4a93abf135d645" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.269258 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5b97647d4d-lzbd6" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.315528 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5b97647d4d-lzbd6"] Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.331415 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5b97647d4d-lzbd6"] Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.497300 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3692143-6629-40ba-984b-187126e4a2ed" path="/var/lib/kubelet/pods/c3692143-6629-40ba-984b-187126e4a2ed/volumes" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.506033 4678 scope.go:117] "RemoveContainer" containerID="d4163c241eb326f5ee5663d607080cbbfd1967af2e9381c39c27f1ceaeebb861" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.656843 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.764453 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2jnz\" (UniqueName: \"kubernetes.io/projected/dea23b4d-e666-4ba6-954d-c07956044128-kube-api-access-h2jnz\") pod \"dea23b4d-e666-4ba6-954d-c07956044128\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.764712 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-scripts\") pod \"dea23b4d-e666-4ba6-954d-c07956044128\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.764739 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-config-data\") pod \"dea23b4d-e666-4ba6-954d-c07956044128\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.764817 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-combined-ca-bundle\") pod \"dea23b4d-e666-4ba6-954d-c07956044128\" (UID: \"dea23b4d-e666-4ba6-954d-c07956044128\") " Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.793108 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-scripts" (OuterVolumeSpecName: "scripts") pod "dea23b4d-e666-4ba6-954d-c07956044128" (UID: "dea23b4d-e666-4ba6-954d-c07956044128"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.793181 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dea23b4d-e666-4ba6-954d-c07956044128-kube-api-access-h2jnz" (OuterVolumeSpecName: "kube-api-access-h2jnz") pod "dea23b4d-e666-4ba6-954d-c07956044128" (UID: "dea23b4d-e666-4ba6-954d-c07956044128"). InnerVolumeSpecName "kube-api-access-h2jnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.807613 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dea23b4d-e666-4ba6-954d-c07956044128" (UID: "dea23b4d-e666-4ba6-954d-c07956044128"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.867832 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.867874 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.867887 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2jnz\" (UniqueName: \"kubernetes.io/projected/dea23b4d-e666-4ba6-954d-c07956044128-kube-api-access-h2jnz\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.877014 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-config-data" (OuterVolumeSpecName: "config-data") pod "dea23b4d-e666-4ba6-954d-c07956044128" (UID: "dea23b4d-e666-4ba6-954d-c07956044128"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:35 crc kubenswrapper[4678]: I1206 10:58:35.969029 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dea23b4d-e666-4ba6-954d-c07956044128-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.294726 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-thzqq" event={"ID":"dea23b4d-e666-4ba6-954d-c07956044128","Type":"ContainerDied","Data":"eec566da813102137aa2e71afe864e3d61ddba7a6731aed8ec151caa330e055a"} Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.294768 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eec566da813102137aa2e71afe864e3d61ddba7a6731aed8ec151caa330e055a" Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.294830 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-thzqq" Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.445845 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.446071 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-log" containerID="cri-o://9f1c3e4b8b048fa251aca2d0e8c0d3446318619e9e59455068c5eafc26e29dda" gracePeriod=30 Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.446171 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-api" containerID="cri-o://e32906004e2a26992632358a2bd005f0d28afb3ddec91dc760ea47752c1825a8" gracePeriod=30 Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.462795 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.463338 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c7ff6af4-3ad5-498d-8d54-39105ba9f01d" containerName="nova-scheduler-scheduler" containerID="cri-o://15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" gracePeriod=30 Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.482039 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.482261 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-log" containerID="cri-o://f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e" gracePeriod=30 Dec 06 10:58:36 crc kubenswrapper[4678]: I1206 10:58:36.482326 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-metadata" containerID="cri-o://ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85" gracePeriod=30 Dec 06 10:58:36 crc kubenswrapper[4678]: E1206 10:58:36.492800 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 10:58:36 crc kubenswrapper[4678]: E1206 10:58:36.498378 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 10:58:36 crc kubenswrapper[4678]: E1206 10:58:36.500324 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 10:58:36 crc kubenswrapper[4678]: E1206 10:58:36.500382 4678 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c7ff6af4-3ad5-498d-8d54-39105ba9f01d" containerName="nova-scheduler-scheduler" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.194540 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.294083 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqlbk\" (UniqueName: \"kubernetes.io/projected/298ed19a-5247-4a42-ae20-a00a511db3ba-kube-api-access-zqlbk\") pod \"298ed19a-5247-4a42-ae20-a00a511db3ba\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.294145 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/298ed19a-5247-4a42-ae20-a00a511db3ba-logs\") pod \"298ed19a-5247-4a42-ae20-a00a511db3ba\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.294226 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-config-data\") pod \"298ed19a-5247-4a42-ae20-a00a511db3ba\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.294333 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-nova-metadata-tls-certs\") pod \"298ed19a-5247-4a42-ae20-a00a511db3ba\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.294387 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-combined-ca-bundle\") pod \"298ed19a-5247-4a42-ae20-a00a511db3ba\" (UID: \"298ed19a-5247-4a42-ae20-a00a511db3ba\") " Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.296130 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/298ed19a-5247-4a42-ae20-a00a511db3ba-logs" (OuterVolumeSpecName: "logs") pod "298ed19a-5247-4a42-ae20-a00a511db3ba" (UID: "298ed19a-5247-4a42-ae20-a00a511db3ba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.312328 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/298ed19a-5247-4a42-ae20-a00a511db3ba-kube-api-access-zqlbk" (OuterVolumeSpecName: "kube-api-access-zqlbk") pod "298ed19a-5247-4a42-ae20-a00a511db3ba" (UID: "298ed19a-5247-4a42-ae20-a00a511db3ba"). InnerVolumeSpecName "kube-api-access-zqlbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.339641 4678 generic.go:334] "Generic (PLEG): container finished" podID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerID="ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85" exitCode=0 Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.339670 4678 generic.go:334] "Generic (PLEG): container finished" podID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerID="f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e" exitCode=143 Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.339702 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.339724 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"298ed19a-5247-4a42-ae20-a00a511db3ba","Type":"ContainerDied","Data":"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85"} Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.339755 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"298ed19a-5247-4a42-ae20-a00a511db3ba","Type":"ContainerDied","Data":"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e"} Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.339765 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"298ed19a-5247-4a42-ae20-a00a511db3ba","Type":"ContainerDied","Data":"de96ff89bb591282200b69e221b91fa59912be1306b4c9b23e5f19109b4caa33"} Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.339778 4678 scope.go:117] "RemoveContainer" containerID="ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.346280 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5031a03-e205-42a3-8c23-6911af163988" containerID="9f1c3e4b8b048fa251aca2d0e8c0d3446318619e9e59455068c5eafc26e29dda" exitCode=143 Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.347050 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5031a03-e205-42a3-8c23-6911af163988","Type":"ContainerDied","Data":"9f1c3e4b8b048fa251aca2d0e8c0d3446318619e9e59455068c5eafc26e29dda"} Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.349637 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "298ed19a-5247-4a42-ae20-a00a511db3ba" (UID: "298ed19a-5247-4a42-ae20-a00a511db3ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.349744 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-config-data" (OuterVolumeSpecName: "config-data") pod "298ed19a-5247-4a42-ae20-a00a511db3ba" (UID: "298ed19a-5247-4a42-ae20-a00a511db3ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.359568 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "298ed19a-5247-4a42-ae20-a00a511db3ba" (UID: "298ed19a-5247-4a42-ae20-a00a511db3ba"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.396570 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.396625 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqlbk\" (UniqueName: \"kubernetes.io/projected/298ed19a-5247-4a42-ae20-a00a511db3ba-kube-api-access-zqlbk\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.396637 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/298ed19a-5247-4a42-ae20-a00a511db3ba-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.396655 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.396665 4678 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/298ed19a-5247-4a42-ae20-a00a511db3ba-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.433076 4678 scope.go:117] "RemoveContainer" containerID="f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.454515 4678 scope.go:117] "RemoveContainer" containerID="ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.456508 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85\": container with ID starting with ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85 not found: ID does not exist" containerID="ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.456546 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85"} err="failed to get container status \"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85\": rpc error: code = NotFound desc = could not find container \"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85\": container with ID starting with ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85 not found: ID does not exist" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.456568 4678 scope.go:117] "RemoveContainer" containerID="f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.456915 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e\": container with ID starting with f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e not found: ID does not exist" containerID="f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.456939 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e"} err="failed to get container status \"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e\": rpc error: code = NotFound desc = could not find container \"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e\": container with ID starting with f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e not found: ID does not exist" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.456953 4678 scope.go:117] "RemoveContainer" containerID="ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.460554 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85"} err="failed to get container status \"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85\": rpc error: code = NotFound desc = could not find container \"ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85\": container with ID starting with ad6ae434976e995615cadad095e1821e5305718bee149cc9f8a24b2900659a85 not found: ID does not exist" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.460573 4678 scope.go:117] "RemoveContainer" containerID="f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.460846 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e"} err="failed to get container status \"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e\": rpc error: code = NotFound desc = could not find container \"f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e\": container with ID starting with f349c53f90d6c423d9d26c51120f5b0fb18f22b26230b163b17146508003cb5e not found: ID does not exist" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.674275 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.694842 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706193 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706724 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-log" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706747 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-log" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706758 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dea23b4d-e666-4ba6-954d-c07956044128" containerName="nova-manage" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706766 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dea23b4d-e666-4ba6-954d-c07956044128" containerName="nova-manage" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706783 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-metadata" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706791 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-metadata" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706807 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerName="init" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706815 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerName="init" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706825 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706832 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706845 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon-log" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706852 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon-log" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706867 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerName="dnsmasq-dns" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706874 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerName="dnsmasq-dns" Dec 06 10:58:37 crc kubenswrapper[4678]: E1206 10:58:37.706884 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.706893 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.707128 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.707145 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1e07f0-3034-47d2-9143-82058df6d73c" containerName="dnsmasq-dns" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.707162 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.707177 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dea23b4d-e666-4ba6-954d-c07956044128" containerName="nova-manage" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.707191 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-metadata" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.707207 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" containerName="nova-metadata-log" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.707221 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3692143-6629-40ba-984b-187126e4a2ed" containerName="horizon-log" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.708460 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.710463 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.713441 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.732636 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.804210 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.804333 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.804375 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fb98dd3-6574-44b5-a779-4c8c9a002637-logs\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.804403 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttzkb\" (UniqueName: \"kubernetes.io/projected/6fb98dd3-6574-44b5-a779-4c8c9a002637-kube-api-access-ttzkb\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.804431 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-config-data\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.906396 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.906943 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.906974 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fb98dd3-6574-44b5-a779-4c8c9a002637-logs\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.907022 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttzkb\" (UniqueName: \"kubernetes.io/projected/6fb98dd3-6574-44b5-a779-4c8c9a002637-kube-api-access-ttzkb\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.907044 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-config-data\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.907522 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fb98dd3-6574-44b5-a779-4c8c9a002637-logs\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.910202 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.910651 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.911629 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-config-data\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:37 crc kubenswrapper[4678]: I1206 10:58:37.923641 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttzkb\" (UniqueName: \"kubernetes.io/projected/6fb98dd3-6574-44b5-a779-4c8c9a002637-kube-api-access-ttzkb\") pod \"nova-metadata-0\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " pod="openstack/nova-metadata-0" Dec 06 10:58:38 crc kubenswrapper[4678]: I1206 10:58:38.033429 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:58:38 crc kubenswrapper[4678]: I1206 10:58:38.359273 4678 generic.go:334] "Generic (PLEG): container finished" podID="02942adf-5e19-482e-a76a-16d40ec52416" containerID="0b127b3a998bc68c192239209fdbf6fee3765fd1f2d5bb7a1fd29011bbf759b7" exitCode=0 Dec 06 10:58:38 crc kubenswrapper[4678]: I1206 10:58:38.359318 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-df7h2" event={"ID":"02942adf-5e19-482e-a76a-16d40ec52416","Type":"ContainerDied","Data":"0b127b3a998bc68c192239209fdbf6fee3765fd1f2d5bb7a1fd29011bbf759b7"} Dec 06 10:58:38 crc kubenswrapper[4678]: I1206 10:58:38.542451 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.079010 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.079462 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="6e0fe084-357f-4b1e-bf7a-bf0413e830b1" containerName="kube-state-metrics" containerID="cri-o://934f5ddaf048b192feb57a1ea23b3ae827e983513df5661e44fa5b2a02a46acf" gracePeriod=30 Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.390513 4678 generic.go:334] "Generic (PLEG): container finished" podID="6e0fe084-357f-4b1e-bf7a-bf0413e830b1" containerID="934f5ddaf048b192feb57a1ea23b3ae827e983513df5661e44fa5b2a02a46acf" exitCode=2 Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.390622 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6e0fe084-357f-4b1e-bf7a-bf0413e830b1","Type":"ContainerDied","Data":"934f5ddaf048b192feb57a1ea23b3ae827e983513df5661e44fa5b2a02a46acf"} Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.404995 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6fb98dd3-6574-44b5-a779-4c8c9a002637","Type":"ContainerStarted","Data":"014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67"} Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.405027 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6fb98dd3-6574-44b5-a779-4c8c9a002637","Type":"ContainerStarted","Data":"a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b"} Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.405038 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6fb98dd3-6574-44b5-a779-4c8c9a002637","Type":"ContainerStarted","Data":"1d328e629e6e6e419fe44ad8333365d831322e75c4a619a7a2e0630c3a865087"} Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.427867 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4278492 podStartE2EDuration="2.4278492s" podCreationTimestamp="2025-12-06 10:58:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:39.427702285 +0000 UTC m=+1324.271133724" watchObservedRunningTime="2025-12-06 10:58:39.4278492 +0000 UTC m=+1324.271280639" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.497018 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="298ed19a-5247-4a42-ae20-a00a511db3ba" path="/var/lib/kubelet/pods/298ed19a-5247-4a42-ae20-a00a511db3ba/volumes" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.604212 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.639425 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6bln\" (UniqueName: \"kubernetes.io/projected/6e0fe084-357f-4b1e-bf7a-bf0413e830b1-kube-api-access-l6bln\") pod \"6e0fe084-357f-4b1e-bf7a-bf0413e830b1\" (UID: \"6e0fe084-357f-4b1e-bf7a-bf0413e830b1\") " Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.646404 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e0fe084-357f-4b1e-bf7a-bf0413e830b1-kube-api-access-l6bln" (OuterVolumeSpecName: "kube-api-access-l6bln") pod "6e0fe084-357f-4b1e-bf7a-bf0413e830b1" (UID: "6e0fe084-357f-4b1e-bf7a-bf0413e830b1"). InnerVolumeSpecName "kube-api-access-l6bln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.741336 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6bln\" (UniqueName: \"kubernetes.io/projected/6e0fe084-357f-4b1e-bf7a-bf0413e830b1-kube-api-access-l6bln\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.759234 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.845178 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8jhj\" (UniqueName: \"kubernetes.io/projected/02942adf-5e19-482e-a76a-16d40ec52416-kube-api-access-s8jhj\") pod \"02942adf-5e19-482e-a76a-16d40ec52416\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.845279 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-config-data\") pod \"02942adf-5e19-482e-a76a-16d40ec52416\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.845356 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-combined-ca-bundle\") pod \"02942adf-5e19-482e-a76a-16d40ec52416\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.845479 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-scripts\") pod \"02942adf-5e19-482e-a76a-16d40ec52416\" (UID: \"02942adf-5e19-482e-a76a-16d40ec52416\") " Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.850805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02942adf-5e19-482e-a76a-16d40ec52416-kube-api-access-s8jhj" (OuterVolumeSpecName: "kube-api-access-s8jhj") pod "02942adf-5e19-482e-a76a-16d40ec52416" (UID: "02942adf-5e19-482e-a76a-16d40ec52416"). InnerVolumeSpecName "kube-api-access-s8jhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.856101 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-scripts" (OuterVolumeSpecName: "scripts") pod "02942adf-5e19-482e-a76a-16d40ec52416" (UID: "02942adf-5e19-482e-a76a-16d40ec52416"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.904712 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-config-data" (OuterVolumeSpecName: "config-data") pod "02942adf-5e19-482e-a76a-16d40ec52416" (UID: "02942adf-5e19-482e-a76a-16d40ec52416"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.906725 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02942adf-5e19-482e-a76a-16d40ec52416" (UID: "02942adf-5e19-482e-a76a-16d40ec52416"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.947906 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.947940 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.947950 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8jhj\" (UniqueName: \"kubernetes.io/projected/02942adf-5e19-482e-a76a-16d40ec52416-kube-api-access-s8jhj\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:39 crc kubenswrapper[4678]: I1206 10:58:39.947958 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02942adf-5e19-482e-a76a-16d40ec52416-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.280776 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.354308 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-combined-ca-bundle\") pod \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.354613 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-config-data\") pod \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.355248 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qblxl\" (UniqueName: \"kubernetes.io/projected/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-kube-api-access-qblxl\") pod \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\" (UID: \"c7ff6af4-3ad5-498d-8d54-39105ba9f01d\") " Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.361208 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-kube-api-access-qblxl" (OuterVolumeSpecName: "kube-api-access-qblxl") pod "c7ff6af4-3ad5-498d-8d54-39105ba9f01d" (UID: "c7ff6af4-3ad5-498d-8d54-39105ba9f01d"). InnerVolumeSpecName "kube-api-access-qblxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.393640 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-config-data" (OuterVolumeSpecName: "config-data") pod "c7ff6af4-3ad5-498d-8d54-39105ba9f01d" (UID: "c7ff6af4-3ad5-498d-8d54-39105ba9f01d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.433531 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c7ff6af4-3ad5-498d-8d54-39105ba9f01d" (UID: "c7ff6af4-3ad5-498d-8d54-39105ba9f01d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.434927 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6e0fe084-357f-4b1e-bf7a-bf0413e830b1","Type":"ContainerDied","Data":"3c99e6479bd2aab8cd7606325e03c75a936d6d9fbbadb37b8a942c909b747445"} Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.434983 4678 scope.go:117] "RemoveContainer" containerID="934f5ddaf048b192feb57a1ea23b3ae827e983513df5661e44fa5b2a02a46acf" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.435122 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.460884 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.460929 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.460941 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qblxl\" (UniqueName: \"kubernetes.io/projected/c7ff6af4-3ad5-498d-8d54-39105ba9f01d-kube-api-access-qblxl\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.461074 4678 generic.go:334] "Generic (PLEG): container finished" podID="e5031a03-e205-42a3-8c23-6911af163988" containerID="e32906004e2a26992632358a2bd005f0d28afb3ddec91dc760ea47752c1825a8" exitCode=0 Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.461158 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5031a03-e205-42a3-8c23-6911af163988","Type":"ContainerDied","Data":"e32906004e2a26992632358a2bd005f0d28afb3ddec91dc760ea47752c1825a8"} Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.470143 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-df7h2" event={"ID":"02942adf-5e19-482e-a76a-16d40ec52416","Type":"ContainerDied","Data":"c47f47fa8fd60ac74b624fbee274c756e3adff47c8f2f327b22a80ae4c5b4b76"} Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.470187 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c47f47fa8fd60ac74b624fbee274c756e3adff47c8f2f327b22a80ae4c5b4b76" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.470284 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-df7h2" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.474628 4678 generic.go:334] "Generic (PLEG): container finished" podID="c7ff6af4-3ad5-498d-8d54-39105ba9f01d" containerID="15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" exitCode=0 Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.479155 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.479309 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c7ff6af4-3ad5-498d-8d54-39105ba9f01d","Type":"ContainerDied","Data":"15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69"} Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.479345 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c7ff6af4-3ad5-498d-8d54-39105ba9f01d","Type":"ContainerDied","Data":"5142ba13cccf95873ebac3a9cea310d0a922c442c39bbbb283bdd93d832ce0bf"} Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.518787 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: E1206 10:58:40.519591 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ff6af4-3ad5-498d-8d54-39105ba9f01d" containerName="nova-scheduler-scheduler" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.519609 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ff6af4-3ad5-498d-8d54-39105ba9f01d" containerName="nova-scheduler-scheduler" Dec 06 10:58:40 crc kubenswrapper[4678]: E1206 10:58:40.519628 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e0fe084-357f-4b1e-bf7a-bf0413e830b1" containerName="kube-state-metrics" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.519634 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e0fe084-357f-4b1e-bf7a-bf0413e830b1" containerName="kube-state-metrics" Dec 06 10:58:40 crc kubenswrapper[4678]: E1206 10:58:40.519657 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02942adf-5e19-482e-a76a-16d40ec52416" containerName="nova-cell1-conductor-db-sync" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.519664 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="02942adf-5e19-482e-a76a-16d40ec52416" containerName="nova-cell1-conductor-db-sync" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.520055 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="02942adf-5e19-482e-a76a-16d40ec52416" containerName="nova-cell1-conductor-db-sync" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.520083 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e0fe084-357f-4b1e-bf7a-bf0413e830b1" containerName="kube-state-metrics" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.520097 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7ff6af4-3ad5-498d-8d54-39105ba9f01d" containerName="nova-scheduler-scheduler" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.521065 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.528125 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.533901 4678 scope.go:117] "RemoveContainer" containerID="15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.534546 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.548561 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.569900 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8srj\" (UniqueName: \"kubernetes.io/projected/ed1b2724-cbb5-4589-9622-6ff20ab1f180-kube-api-access-b8srj\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.569952 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed1b2724-cbb5-4589-9622-6ff20ab1f180-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.570057 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1b2724-cbb5-4589-9622-6ff20ab1f180-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.589432 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.589480 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.606293 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: E1206 10:58:40.607414 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-api" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.607513 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-api" Dec 06 10:58:40 crc kubenswrapper[4678]: E1206 10:58:40.607612 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-log" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.607668 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-log" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.608198 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-log" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.608231 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5031a03-e205-42a3-8c23-6911af163988" containerName="nova-api-api" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.609364 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.609475 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.619164 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.619346 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.640229 4678 scope.go:117] "RemoveContainer" containerID="15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" Dec 06 10:58:40 crc kubenswrapper[4678]: E1206 10:58:40.646893 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69\": container with ID starting with 15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69 not found: ID does not exist" containerID="15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.646941 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69"} err="failed to get container status \"15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69\": rpc error: code = NotFound desc = could not find container \"15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69\": container with ID starting with 15d53c217a43a1cdab4476b88ecc3aac40770537118cd94514513440d599bb69 not found: ID does not exist" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.680115 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.699416 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.704655 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5031a03-e205-42a3-8c23-6911af163988-logs\") pod \"e5031a03-e205-42a3-8c23-6911af163988\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.704787 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-config-data\") pod \"e5031a03-e205-42a3-8c23-6911af163988\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.704842 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqd9r\" (UniqueName: \"kubernetes.io/projected/e5031a03-e205-42a3-8c23-6911af163988-kube-api-access-hqd9r\") pod \"e5031a03-e205-42a3-8c23-6911af163988\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.704878 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-combined-ca-bundle\") pod \"e5031a03-e205-42a3-8c23-6911af163988\" (UID: \"e5031a03-e205-42a3-8c23-6911af163988\") " Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705058 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5031a03-e205-42a3-8c23-6911af163988-logs" (OuterVolumeSpecName: "logs") pod "e5031a03-e205-42a3-8c23-6911af163988" (UID: "e5031a03-e205-42a3-8c23-6911af163988"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705127 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705165 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8srj\" (UniqueName: \"kubernetes.io/projected/ed1b2724-cbb5-4589-9622-6ff20ab1f180-kube-api-access-b8srj\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705185 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed1b2724-cbb5-4589-9622-6ff20ab1f180-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705270 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1b2724-cbb5-4589-9622-6ff20ab1f180-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705296 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705325 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd48x\" (UniqueName: \"kubernetes.io/projected/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-api-access-bd48x\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705356 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.705445 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5031a03-e205-42a3-8c23-6911af163988-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.709059 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5031a03-e205-42a3-8c23-6911af163988-kube-api-access-hqd9r" (OuterVolumeSpecName: "kube-api-access-hqd9r") pod "e5031a03-e205-42a3-8c23-6911af163988" (UID: "e5031a03-e205-42a3-8c23-6911af163988"). InnerVolumeSpecName "kube-api-access-hqd9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.710360 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1b2724-cbb5-4589-9622-6ff20ab1f180-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.710390 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed1b2724-cbb5-4589-9622-6ff20ab1f180-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.723892 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8srj\" (UniqueName: \"kubernetes.io/projected/ed1b2724-cbb5-4589-9622-6ff20ab1f180-kube-api-access-b8srj\") pod \"nova-cell1-conductor-0\" (UID: \"ed1b2724-cbb5-4589-9622-6ff20ab1f180\") " pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.728666 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.730062 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.732621 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.734961 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.738540 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5031a03-e205-42a3-8c23-6911af163988" (UID: "e5031a03-e205-42a3-8c23-6911af163988"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.749247 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-config-data" (OuterVolumeSpecName: "config-data") pod "e5031a03-e205-42a3-8c23-6911af163988" (UID: "e5031a03-e205-42a3-8c23-6911af163988"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.806851 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807138 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd48x\" (UniqueName: \"kubernetes.io/projected/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-api-access-bd48x\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807165 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807193 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807236 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807260 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbfsb\" (UniqueName: \"kubernetes.io/projected/2c3a9d70-4738-4460-b779-5eebd9a06186-kube-api-access-kbfsb\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807345 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-config-data\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807429 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807440 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqd9r\" (UniqueName: \"kubernetes.io/projected/e5031a03-e205-42a3-8c23-6911af163988-kube-api-access-hqd9r\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.807450 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5031a03-e205-42a3-8c23-6911af163988-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.812136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.812894 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.813248 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.822708 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd48x\" (UniqueName: \"kubernetes.io/projected/8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3-kube-api-access-bd48x\") pod \"kube-state-metrics-0\" (UID: \"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3\") " pod="openstack/kube-state-metrics-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.909546 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.909810 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbfsb\" (UniqueName: \"kubernetes.io/projected/2c3a9d70-4738-4460-b779-5eebd9a06186-kube-api-access-kbfsb\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.909935 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-config-data\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.913436 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.917034 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-config-data\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.925329 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbfsb\" (UniqueName: \"kubernetes.io/projected/2c3a9d70-4738-4460-b779-5eebd9a06186-kube-api-access-kbfsb\") pod \"nova-scheduler-0\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " pod="openstack/nova-scheduler-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.961913 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:40 crc kubenswrapper[4678]: I1206 10:58:40.977698 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.057030 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.493075 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e0fe084-357f-4b1e-bf7a-bf0413e830b1" path="/var/lib/kubelet/pods/6e0fe084-357f-4b1e-bf7a-bf0413e830b1/volumes" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.495439 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.496597 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7ff6af4-3ad5-498d-8d54-39105ba9f01d" path="/var/lib/kubelet/pods/c7ff6af4-3ad5-498d-8d54-39105ba9f01d/volumes" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.497893 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5031a03-e205-42a3-8c23-6911af163988","Type":"ContainerDied","Data":"ece93022be59381794959560fa685f54a6afd1fb3907943e7b9ed077be66b6a2"} Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.497931 4678 scope.go:117] "RemoveContainer" containerID="e32906004e2a26992632358a2bd005f0d28afb3ddec91dc760ea47752c1825a8" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.539100 4678 scope.go:117] "RemoveContainer" containerID="9f1c3e4b8b048fa251aca2d0e8c0d3446318619e9e59455068c5eafc26e29dda" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.556465 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.588555 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.619703 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.619777 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.621402 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.624562 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.652371 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.706164 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 06 10:58:41 crc kubenswrapper[4678]: W1206 10:58:41.710284 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c13e227_c6eb_46a3_94a1_e9a3f3d7a6c3.slice/crio-e09b7596b78c04b0abaac90b935daba5e70c61907ff65744c52995def955d00d WatchSource:0}: Error finding container e09b7596b78c04b0abaac90b935daba5e70c61907ff65744c52995def955d00d: Status 404 returned error can't find the container with id e09b7596b78c04b0abaac90b935daba5e70c61907ff65744c52995def955d00d Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.726872 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-logs\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.727275 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.727566 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64h74\" (UniqueName: \"kubernetes.io/projected/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-kube-api-access-64h74\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.727792 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-config-data\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.800356 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.829866 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-config-data\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.830772 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-logs\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.830836 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.831673 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-logs\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.831860 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64h74\" (UniqueName: \"kubernetes.io/projected/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-kube-api-access-64h74\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.834927 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-config-data\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.838755 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:41 crc kubenswrapper[4678]: I1206 10:58:41.857065 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64h74\" (UniqueName: \"kubernetes.io/projected/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-kube-api-access-64h74\") pod \"nova-api-0\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " pod="openstack/nova-api-0" Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:41.999956 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.034777 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.035278 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-central-agent" containerID="cri-o://39f21ec98e9fe3ae2eab904f8be184a9564a0a314cab859458f98488baaa546c" gracePeriod=30 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.035410 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="sg-core" containerID="cri-o://bf57c41d9ca016dcaabdcc5c1cd2c2cb475ef019c6808a1d5e7922a14c7e17c1" gracePeriod=30 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.035428 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-notification-agent" containerID="cri-o://a87ea3eb2113bbda1f5cfac425fb57b3cabc6c6a501235c03e2d09c5884b2983" gracePeriod=30 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.035788 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="proxy-httpd" containerID="cri-o://cc7844f589adf8baeaac0a953c9c63afb6b341d6f73a50cefd4efa2d189be859" gracePeriod=30 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.616906 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.624870 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3","Type":"ContainerStarted","Data":"071be172f5c5e0161f0b31d699a9518ce56cd09e52fdffebd4f93ac073e44357"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.624929 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3","Type":"ContainerStarted","Data":"e09b7596b78c04b0abaac90b935daba5e70c61907ff65744c52995def955d00d"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.625220 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 06 10:58:42 crc kubenswrapper[4678]: W1206 10:58:42.645172 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83bc2dda_e485_4e1a_aa9e_87ea72dfbdd6.slice/crio-32d9b87392cdb68ee668946ccc2298f4ab3729a55f73be8ea7e346295d79e8a8 WatchSource:0}: Error finding container 32d9b87392cdb68ee668946ccc2298f4ab3729a55f73be8ea7e346295d79e8a8: Status 404 returned error can't find the container with id 32d9b87392cdb68ee668946ccc2298f4ab3729a55f73be8ea7e346295d79e8a8 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.645476 4678 generic.go:334] "Generic (PLEG): container finished" podID="1c900c1e-448b-4e4f-abec-a934c6387745" containerID="cc7844f589adf8baeaac0a953c9c63afb6b341d6f73a50cefd4efa2d189be859" exitCode=0 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.645541 4678 generic.go:334] "Generic (PLEG): container finished" podID="1c900c1e-448b-4e4f-abec-a934c6387745" containerID="bf57c41d9ca016dcaabdcc5c1cd2c2cb475ef019c6808a1d5e7922a14c7e17c1" exitCode=2 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.645549 4678 generic.go:334] "Generic (PLEG): container finished" podID="1c900c1e-448b-4e4f-abec-a934c6387745" containerID="39f21ec98e9fe3ae2eab904f8be184a9564a0a314cab859458f98488baaa546c" exitCode=0 Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.645588 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerDied","Data":"cc7844f589adf8baeaac0a953c9c63afb6b341d6f73a50cefd4efa2d189be859"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.645614 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerDied","Data":"bf57c41d9ca016dcaabdcc5c1cd2c2cb475ef019c6808a1d5e7922a14c7e17c1"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.645623 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerDied","Data":"39f21ec98e9fe3ae2eab904f8be184a9564a0a314cab859458f98488baaa546c"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.659330 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ed1b2724-cbb5-4589-9622-6ff20ab1f180","Type":"ContainerStarted","Data":"7cc26b0e84a4f2f231fdf0b930567e9779f06f717e521ec3bbe3d0559521b7b6"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.659367 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ed1b2724-cbb5-4589-9622-6ff20ab1f180","Type":"ContainerStarted","Data":"16872ec405810e6d46c16ac6107458be8a95e429ebca8169496c18eb7d6f900a"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.660270 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.662955 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.260573718 podStartE2EDuration="2.662933505s" podCreationTimestamp="2025-12-06 10:58:40 +0000 UTC" firstStartedPulling="2025-12-06 10:58:41.712772235 +0000 UTC m=+1326.556203674" lastFinishedPulling="2025-12-06 10:58:42.115132022 +0000 UTC m=+1326.958563461" observedRunningTime="2025-12-06 10:58:42.650328297 +0000 UTC m=+1327.493759736" watchObservedRunningTime="2025-12-06 10:58:42.662933505 +0000 UTC m=+1327.506364944" Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.677016 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c3a9d70-4738-4460-b779-5eebd9a06186","Type":"ContainerStarted","Data":"518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.677250 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c3a9d70-4738-4460-b779-5eebd9a06186","Type":"ContainerStarted","Data":"ba15f794fe9537757ddb1cf79bb6f7a7d7e332c8bf769e505f851a66e64c1529"} Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.717769 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.717753053 podStartE2EDuration="2.717753053s" podCreationTimestamp="2025-12-06 10:58:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:42.714051597 +0000 UTC m=+1327.557483036" watchObservedRunningTime="2025-12-06 10:58:42.717753053 +0000 UTC m=+1327.561184492" Dec 06 10:58:42 crc kubenswrapper[4678]: I1206 10:58:42.722897 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.722881805 podStartE2EDuration="2.722881805s" podCreationTimestamp="2025-12-06 10:58:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:42.691469145 +0000 UTC m=+1327.534900584" watchObservedRunningTime="2025-12-06 10:58:42.722881805 +0000 UTC m=+1327.566313244" Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.035105 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.035354 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.487313 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5031a03-e205-42a3-8c23-6911af163988" path="/var/lib/kubelet/pods/e5031a03-e205-42a3-8c23-6911af163988/volumes" Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.697831 4678 generic.go:334] "Generic (PLEG): container finished" podID="1c900c1e-448b-4e4f-abec-a934c6387745" containerID="a87ea3eb2113bbda1f5cfac425fb57b3cabc6c6a501235c03e2d09c5884b2983" exitCode=0 Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.697901 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerDied","Data":"a87ea3eb2113bbda1f5cfac425fb57b3cabc6c6a501235c03e2d09c5884b2983"} Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.710272 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6","Type":"ContainerStarted","Data":"2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08"} Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.710328 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6","Type":"ContainerStarted","Data":"28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31"} Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.710344 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6","Type":"ContainerStarted","Data":"32d9b87392cdb68ee668946ccc2298f4ab3729a55f73be8ea7e346295d79e8a8"} Dec 06 10:58:43 crc kubenswrapper[4678]: I1206 10:58:43.743925 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.743902779 podStartE2EDuration="2.743902779s" podCreationTimestamp="2025-12-06 10:58:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:43.743808296 +0000 UTC m=+1328.587239735" watchObservedRunningTime="2025-12-06 10:58:43.743902779 +0000 UTC m=+1328.587334218" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.048299 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.210749 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-run-httpd\") pod \"1c900c1e-448b-4e4f-abec-a934c6387745\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.210836 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-scripts\") pod \"1c900c1e-448b-4e4f-abec-a934c6387745\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.211052 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-config-data\") pod \"1c900c1e-448b-4e4f-abec-a934c6387745\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.211133 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28rr2\" (UniqueName: \"kubernetes.io/projected/1c900c1e-448b-4e4f-abec-a934c6387745-kube-api-access-28rr2\") pod \"1c900c1e-448b-4e4f-abec-a934c6387745\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.211192 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-sg-core-conf-yaml\") pod \"1c900c1e-448b-4e4f-abec-a934c6387745\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.211228 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-log-httpd\") pod \"1c900c1e-448b-4e4f-abec-a934c6387745\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.211226 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1c900c1e-448b-4e4f-abec-a934c6387745" (UID: "1c900c1e-448b-4e4f-abec-a934c6387745"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.211648 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1c900c1e-448b-4e4f-abec-a934c6387745" (UID: "1c900c1e-448b-4e4f-abec-a934c6387745"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.211735 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-combined-ca-bundle\") pod \"1c900c1e-448b-4e4f-abec-a934c6387745\" (UID: \"1c900c1e-448b-4e4f-abec-a934c6387745\") " Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.212626 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.212652 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c900c1e-448b-4e4f-abec-a934c6387745-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.230384 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-scripts" (OuterVolumeSpecName: "scripts") pod "1c900c1e-448b-4e4f-abec-a934c6387745" (UID: "1c900c1e-448b-4e4f-abec-a934c6387745"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.239111 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c900c1e-448b-4e4f-abec-a934c6387745-kube-api-access-28rr2" (OuterVolumeSpecName: "kube-api-access-28rr2") pod "1c900c1e-448b-4e4f-abec-a934c6387745" (UID: "1c900c1e-448b-4e4f-abec-a934c6387745"). InnerVolumeSpecName "kube-api-access-28rr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.242633 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1c900c1e-448b-4e4f-abec-a934c6387745" (UID: "1c900c1e-448b-4e4f-abec-a934c6387745"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.295521 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c900c1e-448b-4e4f-abec-a934c6387745" (UID: "1c900c1e-448b-4e4f-abec-a934c6387745"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.313861 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28rr2\" (UniqueName: \"kubernetes.io/projected/1c900c1e-448b-4e4f-abec-a934c6387745-kube-api-access-28rr2\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.313894 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.313906 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.313915 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.362345 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-config-data" (OuterVolumeSpecName: "config-data") pod "1c900c1e-448b-4e4f-abec-a934c6387745" (UID: "1c900c1e-448b-4e4f-abec-a934c6387745"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.416036 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c900c1e-448b-4e4f-abec-a934c6387745-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.721104 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c900c1e-448b-4e4f-abec-a934c6387745","Type":"ContainerDied","Data":"4d4958a21c78a6c3138ddea5a430e2f4221933764cf0814b66f169368fa23c69"} Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.721132 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.721482 4678 scope.go:117] "RemoveContainer" containerID="cc7844f589adf8baeaac0a953c9c63afb6b341d6f73a50cefd4efa2d189be859" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.757816 4678 scope.go:117] "RemoveContainer" containerID="bf57c41d9ca016dcaabdcc5c1cd2c2cb475ef019c6808a1d5e7922a14c7e17c1" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.764523 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.791830 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.792205 4678 scope.go:117] "RemoveContainer" containerID="a87ea3eb2113bbda1f5cfac425fb57b3cabc6c6a501235c03e2d09c5884b2983" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.819383 4678 scope.go:117] "RemoveContainer" containerID="39f21ec98e9fe3ae2eab904f8be184a9564a0a314cab859458f98488baaa546c" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.828542 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:44 crc kubenswrapper[4678]: E1206 10:58:44.829088 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-notification-agent" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.829111 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-notification-agent" Dec 06 10:58:44 crc kubenswrapper[4678]: E1206 10:58:44.829138 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-central-agent" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.829147 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-central-agent" Dec 06 10:58:44 crc kubenswrapper[4678]: E1206 10:58:44.829177 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="sg-core" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.829187 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="sg-core" Dec 06 10:58:44 crc kubenswrapper[4678]: E1206 10:58:44.829219 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="proxy-httpd" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.829227 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="proxy-httpd" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.831781 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="proxy-httpd" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.831819 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-notification-agent" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.831834 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="ceilometer-central-agent" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.831850 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" containerName="sg-core" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.834736 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.842919 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.843158 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.843345 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.852064 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.924736 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.924791 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.924826 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-scripts\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.924844 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.925037 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-config-data\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.925083 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-log-httpd\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.925276 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-run-httpd\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:44 crc kubenswrapper[4678]: I1206 10:58:44.925352 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-kube-api-access-l26gl\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-config-data\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027135 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-log-httpd\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027201 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-run-httpd\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027236 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-kube-api-access-l26gl\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027317 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027346 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027378 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-scripts\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.027398 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.028286 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-run-httpd\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.028411 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-log-httpd\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.031859 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.032740 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.033890 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.041597 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-scripts\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.045051 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-kube-api-access-l26gl\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.047871 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-config-data\") pod \"ceilometer-0\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.171908 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.488234 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c900c1e-448b-4e4f-abec-a934c6387745" path="/var/lib/kubelet/pods/1c900c1e-448b-4e4f-abec-a934c6387745/volumes" Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.632533 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:58:45 crc kubenswrapper[4678]: W1206 10:58:45.639604 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd8a569d_9a15_44f8_a53b_d2190dcffc6b.slice/crio-8afa1ef167a08240cb1d710fa8e4a60656e005e461147b24048674317a1cd5f5 WatchSource:0}: Error finding container 8afa1ef167a08240cb1d710fa8e4a60656e005e461147b24048674317a1cd5f5: Status 404 returned error can't find the container with id 8afa1ef167a08240cb1d710fa8e4a60656e005e461147b24048674317a1cd5f5 Dec 06 10:58:45 crc kubenswrapper[4678]: I1206 10:58:45.734477 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerStarted","Data":"8afa1ef167a08240cb1d710fa8e4a60656e005e461147b24048674317a1cd5f5"} Dec 06 10:58:46 crc kubenswrapper[4678]: I1206 10:58:46.058405 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 10:58:46 crc kubenswrapper[4678]: I1206 10:58:46.744948 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerStarted","Data":"30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba"} Dec 06 10:58:47 crc kubenswrapper[4678]: I1206 10:58:47.756973 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerStarted","Data":"cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188"} Dec 06 10:58:48 crc kubenswrapper[4678]: I1206 10:58:48.034096 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 10:58:48 crc kubenswrapper[4678]: I1206 10:58:48.034148 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 10:58:48 crc kubenswrapper[4678]: I1206 10:58:48.788069 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerStarted","Data":"418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921"} Dec 06 10:58:49 crc kubenswrapper[4678]: I1206 10:58:49.046718 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:58:49 crc kubenswrapper[4678]: I1206 10:58:49.046731 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.189:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:58:49 crc kubenswrapper[4678]: I1206 10:58:49.798509 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerStarted","Data":"a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179"} Dec 06 10:58:49 crc kubenswrapper[4678]: I1206 10:58:49.798808 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 10:58:49 crc kubenswrapper[4678]: I1206 10:58:49.821892 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.025447089 podStartE2EDuration="5.821866053s" podCreationTimestamp="2025-12-06 10:58:44 +0000 UTC" firstStartedPulling="2025-12-06 10:58:45.642886966 +0000 UTC m=+1330.486318415" lastFinishedPulling="2025-12-06 10:58:49.43930594 +0000 UTC m=+1334.282737379" observedRunningTime="2025-12-06 10:58:49.816567686 +0000 UTC m=+1334.659999135" watchObservedRunningTime="2025-12-06 10:58:49.821866053 +0000 UTC m=+1334.665297492" Dec 06 10:58:50 crc kubenswrapper[4678]: I1206 10:58:50.998083 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 06 10:58:51 crc kubenswrapper[4678]: I1206 10:58:51.027636 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 06 10:58:51 crc kubenswrapper[4678]: I1206 10:58:51.058213 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 10:58:51 crc kubenswrapper[4678]: I1206 10:58:51.142127 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 10:58:51 crc kubenswrapper[4678]: I1206 10:58:51.880142 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 10:58:52 crc kubenswrapper[4678]: I1206 10:58:52.001177 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 10:58:52 crc kubenswrapper[4678]: I1206 10:58:52.001253 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 10:58:53 crc kubenswrapper[4678]: I1206 10:58:53.084836 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:58:53 crc kubenswrapper[4678]: I1206 10:58:53.084837 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.193:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.573198 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.724482 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpz6k\" (UniqueName: \"kubernetes.io/projected/5a9b1e8b-82ef-4abb-aff7-751e89de309f-kube-api-access-qpz6k\") pod \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.725009 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-combined-ca-bundle\") pod \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.725207 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-config-data\") pod \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\" (UID: \"5a9b1e8b-82ef-4abb-aff7-751e89de309f\") " Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.733601 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a9b1e8b-82ef-4abb-aff7-751e89de309f-kube-api-access-qpz6k" (OuterVolumeSpecName: "kube-api-access-qpz6k") pod "5a9b1e8b-82ef-4abb-aff7-751e89de309f" (UID: "5a9b1e8b-82ef-4abb-aff7-751e89de309f"). InnerVolumeSpecName "kube-api-access-qpz6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.754783 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-config-data" (OuterVolumeSpecName: "config-data") pod "5a9b1e8b-82ef-4abb-aff7-751e89de309f" (UID: "5a9b1e8b-82ef-4abb-aff7-751e89de309f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.761133 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a9b1e8b-82ef-4abb-aff7-751e89de309f" (UID: "5a9b1e8b-82ef-4abb-aff7-751e89de309f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.827947 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpz6k\" (UniqueName: \"kubernetes.io/projected/5a9b1e8b-82ef-4abb-aff7-751e89de309f-kube-api-access-qpz6k\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.827977 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.827987 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9b1e8b-82ef-4abb-aff7-751e89de309f-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.896080 4678 generic.go:334] "Generic (PLEG): container finished" podID="5a9b1e8b-82ef-4abb-aff7-751e89de309f" containerID="3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2" exitCode=137 Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.896125 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5a9b1e8b-82ef-4abb-aff7-751e89de309f","Type":"ContainerDied","Data":"3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2"} Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.896152 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"5a9b1e8b-82ef-4abb-aff7-751e89de309f","Type":"ContainerDied","Data":"0d44d8ad2b506e6e56027ad477a79c74a9b78515d303469fa89346264e3888dd"} Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.896169 4678 scope.go:117] "RemoveContainer" containerID="3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.896171 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.942826 4678 scope.go:117] "RemoveContainer" containerID="3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2" Dec 06 10:58:57 crc kubenswrapper[4678]: E1206 10:58:57.943911 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2\": container with ID starting with 3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2 not found: ID does not exist" containerID="3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.943961 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2"} err="failed to get container status \"3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2\": rpc error: code = NotFound desc = could not find container \"3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2\": container with ID starting with 3aaef2b6c6e0b2397829beec407ba9f31c6dfc512f59e798c073c5d759796ff2 not found: ID does not exist" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.945925 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.977891 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.995448 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:57 crc kubenswrapper[4678]: E1206 10:58:57.995952 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a9b1e8b-82ef-4abb-aff7-751e89de309f" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.995998 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a9b1e8b-82ef-4abb-aff7-751e89de309f" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.996244 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a9b1e8b-82ef-4abb-aff7-751e89de309f" containerName="nova-cell1-novncproxy-novncproxy" Dec 06 10:58:57 crc kubenswrapper[4678]: I1206 10:58:57.999251 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.001261 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.001320 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.001586 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.006635 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.040594 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.045958 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.046879 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.134846 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dd5t\" (UniqueName: \"kubernetes.io/projected/21d64141-1313-4582-ac36-a0d0a304f0fb-kube-api-access-5dd5t\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.135026 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.135187 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.135434 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.135604 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.239046 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dd5t\" (UniqueName: \"kubernetes.io/projected/21d64141-1313-4582-ac36-a0d0a304f0fb-kube-api-access-5dd5t\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.239203 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.239309 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.239442 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.239552 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.243839 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.245420 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.246118 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.246664 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/21d64141-1313-4582-ac36-a0d0a304f0fb-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.265997 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dd5t\" (UniqueName: \"kubernetes.io/projected/21d64141-1313-4582-ac36-a0d0a304f0fb-kube-api-access-5dd5t\") pod \"nova-cell1-novncproxy-0\" (UID: \"21d64141-1313-4582-ac36-a0d0a304f0fb\") " pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.322274 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.790633 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.914344 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"21d64141-1313-4582-ac36-a0d0a304f0fb","Type":"ContainerStarted","Data":"f54d9fded014fab733b47bb5757bcbe64d22b86fa39472233728683ddffbfb2b"} Dec 06 10:58:58 crc kubenswrapper[4678]: I1206 10:58:58.920613 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 10:58:59 crc kubenswrapper[4678]: I1206 10:58:59.485414 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a9b1e8b-82ef-4abb-aff7-751e89de309f" path="/var/lib/kubelet/pods/5a9b1e8b-82ef-4abb-aff7-751e89de309f/volumes" Dec 06 10:58:59 crc kubenswrapper[4678]: I1206 10:58:59.505378 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:58:59 crc kubenswrapper[4678]: I1206 10:58:59.505434 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:58:59 crc kubenswrapper[4678]: I1206 10:58:59.923694 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"21d64141-1313-4582-ac36-a0d0a304f0fb","Type":"ContainerStarted","Data":"988e8c2334d4b53529d4d8ff9809e4a556926ca70e6b8531ae8d0e637718a1ef"} Dec 06 10:58:59 crc kubenswrapper[4678]: I1206 10:58:59.955350 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.9553249900000003 podStartE2EDuration="2.95532499s" podCreationTimestamp="2025-12-06 10:58:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:58:59.946268994 +0000 UTC m=+1344.789700443" watchObservedRunningTime="2025-12-06 10:58:59.95532499 +0000 UTC m=+1344.798756449" Dec 06 10:59:02 crc kubenswrapper[4678]: I1206 10:59:02.007095 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 10:59:02 crc kubenswrapper[4678]: I1206 10:59:02.008379 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 10:59:02 crc kubenswrapper[4678]: I1206 10:59:02.010139 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 10:59:02 crc kubenswrapper[4678]: I1206 10:59:02.013598 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 10:59:02 crc kubenswrapper[4678]: I1206 10:59:02.952696 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 10:59:02 crc kubenswrapper[4678]: I1206 10:59:02.956652 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.160579 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-ld2rs"] Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.162873 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.183539 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-ld2rs"] Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.245331 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.245458 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-config\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.245530 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.245557 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.245723 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.245873 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxvbs\" (UniqueName: \"kubernetes.io/projected/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-kube-api-access-gxvbs\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.323428 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.347676 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.347789 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxvbs\" (UniqueName: \"kubernetes.io/projected/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-kube-api-access-gxvbs\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.347841 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.347880 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-config\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.347920 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.347941 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.349093 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.349115 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-config\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.349536 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.349806 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.350712 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.365813 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxvbs\" (UniqueName: \"kubernetes.io/projected/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-kube-api-access-gxvbs\") pod \"dnsmasq-dns-cd5cbd7b9-ld2rs\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:03 crc kubenswrapper[4678]: I1206 10:59:03.511508 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:04 crc kubenswrapper[4678]: I1206 10:59:04.050768 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-ld2rs"] Dec 06 10:59:04 crc kubenswrapper[4678]: I1206 10:59:04.988089 4678 generic.go:334] "Generic (PLEG): container finished" podID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerID="5e84ff4077e11d33c2dc981d1419d793f408cb1c405425671042459174bf04f4" exitCode=0 Dec 06 10:59:04 crc kubenswrapper[4678]: I1206 10:59:04.988189 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" event={"ID":"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1","Type":"ContainerDied","Data":"5e84ff4077e11d33c2dc981d1419d793f408cb1c405425671042459174bf04f4"} Dec 06 10:59:04 crc kubenswrapper[4678]: I1206 10:59:04.989638 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" event={"ID":"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1","Type":"ContainerStarted","Data":"b3b4767b0a2aefef7cf7371efaaef6cf69654509cf13c367f51280fcc889f8c4"} Dec 06 10:59:05 crc kubenswrapper[4678]: I1206 10:59:05.985087 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:59:05 crc kubenswrapper[4678]: I1206 10:59:05.985721 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-central-agent" containerID="cri-o://30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba" gracePeriod=30 Dec 06 10:59:05 crc kubenswrapper[4678]: I1206 10:59:05.985851 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="proxy-httpd" containerID="cri-o://a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179" gracePeriod=30 Dec 06 10:59:05 crc kubenswrapper[4678]: I1206 10:59:05.985939 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-notification-agent" containerID="cri-o://cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188" gracePeriod=30 Dec 06 10:59:05 crc kubenswrapper[4678]: I1206 10:59:05.986076 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="sg-core" containerID="cri-o://418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921" gracePeriod=30 Dec 06 10:59:06 crc kubenswrapper[4678]: I1206 10:59:06.020157 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" event={"ID":"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1","Type":"ContainerStarted","Data":"ea562793da98cd8d0e1a3624820e2b17ae7f6e8fae92eeb224d5f48b247ecf3e"} Dec 06 10:59:06 crc kubenswrapper[4678]: I1206 10:59:06.021620 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:06 crc kubenswrapper[4678]: I1206 10:59:06.031308 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.194:3000/\": EOF" Dec 06 10:59:06 crc kubenswrapper[4678]: I1206 10:59:06.046564 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" podStartSLOduration=3.046547742 podStartE2EDuration="3.046547742s" podCreationTimestamp="2025-12-06 10:59:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:59:06.042411763 +0000 UTC m=+1350.885843202" watchObservedRunningTime="2025-12-06 10:59:06.046547742 +0000 UTC m=+1350.889979181" Dec 06 10:59:06 crc kubenswrapper[4678]: I1206 10:59:06.300308 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:06 crc kubenswrapper[4678]: I1206 10:59:06.300558 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-log" containerID="cri-o://28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31" gracePeriod=30 Dec 06 10:59:06 crc kubenswrapper[4678]: I1206 10:59:06.300701 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-api" containerID="cri-o://2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08" gracePeriod=30 Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.390254 4678 generic.go:334] "Generic (PLEG): container finished" podID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerID="28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31" exitCode=143 Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.390427 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6","Type":"ContainerDied","Data":"28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31"} Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.394318 4678 generic.go:334] "Generic (PLEG): container finished" podID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerID="a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179" exitCode=0 Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.394343 4678 generic.go:334] "Generic (PLEG): container finished" podID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerID="418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921" exitCode=2 Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.394357 4678 generic.go:334] "Generic (PLEG): container finished" podID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerID="30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba" exitCode=0 Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.394355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerDied","Data":"a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179"} Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.394389 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerDied","Data":"418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921"} Dec 06 10:59:07 crc kubenswrapper[4678]: I1206 10:59:07.394415 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerDied","Data":"30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba"} Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.323054 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.343141 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.435530 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.585985 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9wvrw"] Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.587575 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.591618 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.591869 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.596529 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9wvrw"] Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.597359 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pshk6\" (UniqueName: \"kubernetes.io/projected/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-kube-api-access-pshk6\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.597445 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-scripts\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.597476 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-config-data\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.597580 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.698632 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-scripts\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.698672 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-config-data\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.698759 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.698789 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pshk6\" (UniqueName: \"kubernetes.io/projected/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-kube-api-access-pshk6\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.716562 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-scripts\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.719150 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-config-data\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.719956 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pshk6\" (UniqueName: \"kubernetes.io/projected/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-kube-api-access-pshk6\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.726135 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9wvrw\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:08 crc kubenswrapper[4678]: I1206 10:59:08.906108 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.102978 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.206930 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-config-data\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.206981 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-combined-ca-bundle\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.207028 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-ceilometer-tls-certs\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.207156 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-log-httpd\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.207191 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-scripts\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.207253 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-run-httpd\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.207278 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-sg-core-conf-yaml\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.207394 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-kube-api-access-l26gl\") pod \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\" (UID: \"fd8a569d-9a15-44f8-a53b-d2190dcffc6b\") " Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.208023 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.208780 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.214563 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-kube-api-access-l26gl" (OuterVolumeSpecName: "kube-api-access-l26gl") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "kube-api-access-l26gl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.218189 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-scripts" (OuterVolumeSpecName: "scripts") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.259894 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.286638 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.314837 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l26gl\" (UniqueName: \"kubernetes.io/projected/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-kube-api-access-l26gl\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.314874 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.314884 4678 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.314893 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.314904 4678 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.314912 4678 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.331688 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.342646 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-config-data" (OuterVolumeSpecName: "config-data") pod "fd8a569d-9a15-44f8-a53b-d2190dcffc6b" (UID: "fd8a569d-9a15-44f8-a53b-d2190dcffc6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.414542 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.414610 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerDied","Data":"cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188"} Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.414982 4678 scope.go:117] "RemoveContainer" containerID="a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.415148 4678 generic.go:334] "Generic (PLEG): container finished" podID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerID="cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188" exitCode=0 Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.415228 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd8a569d-9a15-44f8-a53b-d2190dcffc6b","Type":"ContainerDied","Data":"8afa1ef167a08240cb1d710fa8e4a60656e005e461147b24048674317a1cd5f5"} Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.416173 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.416430 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd8a569d-9a15-44f8-a53b-d2190dcffc6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.438734 4678 scope.go:117] "RemoveContainer" containerID="418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.463341 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.464726 4678 scope.go:117] "RemoveContainer" containerID="cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.508819 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.511056 4678 scope.go:117] "RemoveContainer" containerID="30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.526105 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9wvrw"] Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.560647 4678 scope.go:117] "RemoveContainer" containerID="a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179" Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.562770 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179\": container with ID starting with a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179 not found: ID does not exist" containerID="a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.562825 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179"} err="failed to get container status \"a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179\": rpc error: code = NotFound desc = could not find container \"a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179\": container with ID starting with a48a06f7f2eb3ddef4dcdfc9628224ff95f4fff20f5542c79409234f10a8e179 not found: ID does not exist" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.562860 4678 scope.go:117] "RemoveContainer" containerID="418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921" Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.566108 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921\": container with ID starting with 418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921 not found: ID does not exist" containerID="418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.566146 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921"} err="failed to get container status \"418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921\": rpc error: code = NotFound desc = could not find container \"418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921\": container with ID starting with 418000b55f0c9eaea351d94da8ace285c89a833b84e2038ccbdf6b67ee765921 not found: ID does not exist" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.566163 4678 scope.go:117] "RemoveContainer" containerID="cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188" Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.568263 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188\": container with ID starting with cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188 not found: ID does not exist" containerID="cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.568290 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188"} err="failed to get container status \"cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188\": rpc error: code = NotFound desc = could not find container \"cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188\": container with ID starting with cb3a47d875cca9d912459fc4ba363f246ce7007f9e560212862988b92eec1188 not found: ID does not exist" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.568304 4678 scope.go:117] "RemoveContainer" containerID="30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba" Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.568782 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba\": container with ID starting with 30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba not found: ID does not exist" containerID="30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.568803 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba"} err="failed to get container status \"30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba\": rpc error: code = NotFound desc = could not find container \"30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba\": container with ID starting with 30c09175e5867123c43639a627795c95216a9081dc414319ab69c058c8c28cba not found: ID does not exist" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.570920 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.571389 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-notification-agent" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571412 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-notification-agent" Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.571438 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-central-agent" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571444 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-central-agent" Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.571455 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="proxy-httpd" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571463 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="proxy-httpd" Dec 06 10:59:09 crc kubenswrapper[4678]: E1206 10:59:09.571479 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="sg-core" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571501 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="sg-core" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571739 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="sg-core" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571775 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-notification-agent" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571795 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="ceilometer-central-agent" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.571811 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" containerName="proxy-httpd" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.580919 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.583256 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.583419 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.583522 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.583761 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.622713 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-scripts\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.622778 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-run-httpd\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.622836 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.622886 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.622913 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-log-httpd\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.622956 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8m25\" (UniqueName: \"kubernetes.io/projected/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-kube-api-access-w8m25\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.622974 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.623003 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-config-data\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.723684 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-run-httpd\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.723781 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.723816 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.723845 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-log-httpd\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.723902 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8m25\" (UniqueName: \"kubernetes.io/projected/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-kube-api-access-w8m25\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.723925 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.723966 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-config-data\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.724009 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-scripts\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.724335 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-run-httpd\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.724571 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-log-httpd\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.731103 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-config-data\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.732996 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.733674 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-scripts\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.735267 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.742508 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.771941 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8m25\" (UniqueName: \"kubernetes.io/projected/5a9c85b9-5f97-4449-ba76-0f81fd00cce1-kube-api-access-w8m25\") pod \"ceilometer-0\" (UID: \"5a9c85b9-5f97-4449-ba76-0f81fd00cce1\") " pod="openstack/ceilometer-0" Dec 06 10:59:09 crc kubenswrapper[4678]: I1206 10:59:09.914957 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.033705 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.242134 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-logs\") pod \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.242501 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-config-data\") pod \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.242540 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-combined-ca-bundle\") pod \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.242722 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64h74\" (UniqueName: \"kubernetes.io/projected/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-kube-api-access-64h74\") pod \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\" (UID: \"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6\") " Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.243510 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-logs" (OuterVolumeSpecName: "logs") pod "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" (UID: "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.244234 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.292111 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-kube-api-access-64h74" (OuterVolumeSpecName: "kube-api-access-64h74") pod "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" (UID: "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6"). InnerVolumeSpecName "kube-api-access-64h74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.344946 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64h74\" (UniqueName: \"kubernetes.io/projected/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-kube-api-access-64h74\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.375147 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-config-data" (OuterVolumeSpecName: "config-data") pod "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" (UID: "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.383609 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" (UID: "83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.446083 4678 generic.go:334] "Generic (PLEG): container finished" podID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerID="2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08" exitCode=0 Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.446151 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6","Type":"ContainerDied","Data":"2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08"} Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.446185 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6","Type":"ContainerDied","Data":"32d9b87392cdb68ee668946ccc2298f4ab3729a55f73be8ea7e346295d79e8a8"} Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.446201 4678 scope.go:117] "RemoveContainer" containerID="2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.446299 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.449253 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.449329 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.485751 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9wvrw" event={"ID":"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46","Type":"ContainerStarted","Data":"4078ef8e9e930b90e52963fd33488744b48575ad03bc0edee8ee104b5cb41795"} Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.485789 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9wvrw" event={"ID":"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46","Type":"ContainerStarted","Data":"9f90c0666324fd16bbaa1ff3fbafde5944780b42e3365d6c14cc05495cf90e73"} Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.503670 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.517141 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.562362 4678 scope.go:117] "RemoveContainer" containerID="28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.568154 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:10 crc kubenswrapper[4678]: E1206 10:59:10.568481 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-log" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.568507 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-log" Dec 06 10:59:10 crc kubenswrapper[4678]: E1206 10:59:10.568533 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-api" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.568539 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-api" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.568702 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-log" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.568726 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" containerName="nova-api-api" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.570393 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.574002 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.574500 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.574593 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.587396 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9wvrw" podStartSLOduration=2.5873703089999998 podStartE2EDuration="2.587370309s" podCreationTimestamp="2025-12-06 10:59:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:59:10.558708036 +0000 UTC m=+1355.402139465" watchObservedRunningTime="2025-12-06 10:59:10.587370309 +0000 UTC m=+1355.430801758" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.638668 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.643614 4678 scope.go:117] "RemoveContainer" containerID="2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08" Dec 06 10:59:10 crc kubenswrapper[4678]: E1206 10:59:10.644858 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08\": container with ID starting with 2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08 not found: ID does not exist" containerID="2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.644924 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08"} err="failed to get container status \"2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08\": rpc error: code = NotFound desc = could not find container \"2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08\": container with ID starting with 2556c37656908d8a5b6d74514274aa35dc6bb14eda4c59b0d2519c26a2afcc08 not found: ID does not exist" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.644956 4678 scope.go:117] "RemoveContainer" containerID="28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31" Dec 06 10:59:10 crc kubenswrapper[4678]: E1206 10:59:10.645296 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31\": container with ID starting with 28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31 not found: ID does not exist" containerID="28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.645314 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31"} err="failed to get container status \"28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31\": rpc error: code = NotFound desc = could not find container \"28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31\": container with ID starting with 28132ff0f3806557cd2698c76f56b1a60a0ee2e8b0128904734556fcd01dba31 not found: ID does not exist" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.746043 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.769659 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.769692 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-config-data\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.769840 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.769891 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79e75e4-bc70-46e9-b808-58a9a371e045-logs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.769964 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9qdt\" (UniqueName: \"kubernetes.io/projected/b79e75e4-bc70-46e9-b808-58a9a371e045-kube-api-access-s9qdt\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.769985 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-public-tls-certs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.871947 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9qdt\" (UniqueName: \"kubernetes.io/projected/b79e75e4-bc70-46e9-b808-58a9a371e045-kube-api-access-s9qdt\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.872018 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-public-tls-certs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.872270 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.872291 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-config-data\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.872348 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.872380 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79e75e4-bc70-46e9-b808-58a9a371e045-logs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.873195 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79e75e4-bc70-46e9-b808-58a9a371e045-logs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.879351 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.881258 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.881857 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-config-data\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.882365 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-public-tls-certs\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.903038 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9qdt\" (UniqueName: \"kubernetes.io/projected/b79e75e4-bc70-46e9-b808-58a9a371e045-kube-api-access-s9qdt\") pod \"nova-api-0\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " pod="openstack/nova-api-0" Dec 06 10:59:10 crc kubenswrapper[4678]: I1206 10:59:10.919328 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:11 crc kubenswrapper[4678]: W1206 10:59:11.432881 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb79e75e4_bc70_46e9_b808_58a9a371e045.slice/crio-e00fc95a020cc4a661fccbbdeee0dfe5a696f9a36984a5bb9ab9637e5c6a28ce WatchSource:0}: Error finding container e00fc95a020cc4a661fccbbdeee0dfe5a696f9a36984a5bb9ab9637e5c6a28ce: Status 404 returned error can't find the container with id e00fc95a020cc4a661fccbbdeee0dfe5a696f9a36984a5bb9ab9637e5c6a28ce Dec 06 10:59:11 crc kubenswrapper[4678]: I1206 10:59:11.434421 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:11 crc kubenswrapper[4678]: I1206 10:59:11.497548 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6" path="/var/lib/kubelet/pods/83bc2dda-e485-4e1a-aa9e-87ea72dfbdd6/volumes" Dec 06 10:59:11 crc kubenswrapper[4678]: I1206 10:59:11.498207 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd8a569d-9a15-44f8-a53b-d2190dcffc6b" path="/var/lib/kubelet/pods/fd8a569d-9a15-44f8-a53b-d2190dcffc6b/volumes" Dec 06 10:59:11 crc kubenswrapper[4678]: I1206 10:59:11.508517 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a9c85b9-5f97-4449-ba76-0f81fd00cce1","Type":"ContainerStarted","Data":"861f6a78bdb5ce404bb9388c6e511a545d57a6c0f52e27274f41419a31865e25"} Dec 06 10:59:11 crc kubenswrapper[4678]: I1206 10:59:11.508564 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a9c85b9-5f97-4449-ba76-0f81fd00cce1","Type":"ContainerStarted","Data":"368f2aaab6a31ea83e52526c4102cf203f3c5454d5b3c75514fd77686080756a"} Dec 06 10:59:11 crc kubenswrapper[4678]: I1206 10:59:11.511440 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b79e75e4-bc70-46e9-b808-58a9a371e045","Type":"ContainerStarted","Data":"e00fc95a020cc4a661fccbbdeee0dfe5a696f9a36984a5bb9ab9637e5c6a28ce"} Dec 06 10:59:12 crc kubenswrapper[4678]: I1206 10:59:12.554674 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b79e75e4-bc70-46e9-b808-58a9a371e045","Type":"ContainerStarted","Data":"435762cba1461d3cd78d31088be4ff9c7976bc87d2b1c212357b5b1d03f7bda0"} Dec 06 10:59:12 crc kubenswrapper[4678]: I1206 10:59:12.555263 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b79e75e4-bc70-46e9-b808-58a9a371e045","Type":"ContainerStarted","Data":"4e4e6974c8af5150f499a772ee857cb6878a20594d12b388de39f7fc6c0be9df"} Dec 06 10:59:12 crc kubenswrapper[4678]: I1206 10:59:12.563710 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a9c85b9-5f97-4449-ba76-0f81fd00cce1","Type":"ContainerStarted","Data":"1ea86a71141741c07bbf98ab277260b9d6042d51be668b776b8063adc56b77f9"} Dec 06 10:59:12 crc kubenswrapper[4678]: I1206 10:59:12.605010 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.604987367 podStartE2EDuration="2.604987367s" podCreationTimestamp="2025-12-06 10:59:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:59:12.579474682 +0000 UTC m=+1357.422906121" watchObservedRunningTime="2025-12-06 10:59:12.604987367 +0000 UTC m=+1357.448418806" Dec 06 10:59:13 crc kubenswrapper[4678]: I1206 10:59:13.513681 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 10:59:13 crc kubenswrapper[4678]: I1206 10:59:13.592443 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a9c85b9-5f97-4449-ba76-0f81fd00cce1","Type":"ContainerStarted","Data":"0ad97b55ee057d3da487eeafa6db3932f740682f5d612d91e02047ea01c0f3ec"} Dec 06 10:59:13 crc kubenswrapper[4678]: I1206 10:59:13.608348 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-9brh9"] Dec 06 10:59:13 crc kubenswrapper[4678]: I1206 10:59:13.608778 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" podUID="69716076-6f0e-4432-91fc-82ae535dd96b" containerName="dnsmasq-dns" containerID="cri-o://e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12" gracePeriod=10 Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.104819 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.245219 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-sb\") pod \"69716076-6f0e-4432-91fc-82ae535dd96b\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.245688 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-swift-storage-0\") pod \"69716076-6f0e-4432-91fc-82ae535dd96b\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.245724 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-config\") pod \"69716076-6f0e-4432-91fc-82ae535dd96b\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.245768 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-nb\") pod \"69716076-6f0e-4432-91fc-82ae535dd96b\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.245786 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9sxx\" (UniqueName: \"kubernetes.io/projected/69716076-6f0e-4432-91fc-82ae535dd96b-kube-api-access-d9sxx\") pod \"69716076-6f0e-4432-91fc-82ae535dd96b\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.245808 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-svc\") pod \"69716076-6f0e-4432-91fc-82ae535dd96b\" (UID: \"69716076-6f0e-4432-91fc-82ae535dd96b\") " Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.253765 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69716076-6f0e-4432-91fc-82ae535dd96b-kube-api-access-d9sxx" (OuterVolumeSpecName: "kube-api-access-d9sxx") pod "69716076-6f0e-4432-91fc-82ae535dd96b" (UID: "69716076-6f0e-4432-91fc-82ae535dd96b"). InnerVolumeSpecName "kube-api-access-d9sxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.325623 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-config" (OuterVolumeSpecName: "config") pod "69716076-6f0e-4432-91fc-82ae535dd96b" (UID: "69716076-6f0e-4432-91fc-82ae535dd96b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.343394 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "69716076-6f0e-4432-91fc-82ae535dd96b" (UID: "69716076-6f0e-4432-91fc-82ae535dd96b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.348546 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "69716076-6f0e-4432-91fc-82ae535dd96b" (UID: "69716076-6f0e-4432-91fc-82ae535dd96b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.349034 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.349065 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-config\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.349075 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.349085 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9sxx\" (UniqueName: \"kubernetes.io/projected/69716076-6f0e-4432-91fc-82ae535dd96b-kube-api-access-d9sxx\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.356821 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69716076-6f0e-4432-91fc-82ae535dd96b" (UID: "69716076-6f0e-4432-91fc-82ae535dd96b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.368016 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "69716076-6f0e-4432-91fc-82ae535dd96b" (UID: "69716076-6f0e-4432-91fc-82ae535dd96b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.451247 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.451454 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/69716076-6f0e-4432-91fc-82ae535dd96b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.603199 4678 generic.go:334] "Generic (PLEG): container finished" podID="69716076-6f0e-4432-91fc-82ae535dd96b" containerID="e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12" exitCode=0 Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.603241 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" event={"ID":"69716076-6f0e-4432-91fc-82ae535dd96b","Type":"ContainerDied","Data":"e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12"} Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.603263 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.603286 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-9brh9" event={"ID":"69716076-6f0e-4432-91fc-82ae535dd96b","Type":"ContainerDied","Data":"fbc11a2bb670858f8c7ab950319dc9f1097fd08a25221c6dda10006aec4747a7"} Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.603304 4678 scope.go:117] "RemoveContainer" containerID="e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.635362 4678 scope.go:117] "RemoveContainer" containerID="a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.637729 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-9brh9"] Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.647713 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-9brh9"] Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.664511 4678 scope.go:117] "RemoveContainer" containerID="e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12" Dec 06 10:59:14 crc kubenswrapper[4678]: E1206 10:59:14.664984 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12\": container with ID starting with e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12 not found: ID does not exist" containerID="e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.665037 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12"} err="failed to get container status \"e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12\": rpc error: code = NotFound desc = could not find container \"e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12\": container with ID starting with e7169dbf225742a2ee4628edcbc5618c0f63aaeef627f54e82c0e05a4c445b12 not found: ID does not exist" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.665072 4678 scope.go:117] "RemoveContainer" containerID="a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe" Dec 06 10:59:14 crc kubenswrapper[4678]: E1206 10:59:14.665454 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe\": container with ID starting with a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe not found: ID does not exist" containerID="a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe" Dec 06 10:59:14 crc kubenswrapper[4678]: I1206 10:59:14.665534 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe"} err="failed to get container status \"a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe\": rpc error: code = NotFound desc = could not find container \"a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe\": container with ID starting with a35d5d1daf75a76aa6b1a551759f8c1755a26a743755b733dc409b48081aa8fe not found: ID does not exist" Dec 06 10:59:15 crc kubenswrapper[4678]: I1206 10:59:15.546034 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69716076-6f0e-4432-91fc-82ae535dd96b" path="/var/lib/kubelet/pods/69716076-6f0e-4432-91fc-82ae535dd96b/volumes" Dec 06 10:59:16 crc kubenswrapper[4678]: I1206 10:59:16.649757 4678 generic.go:334] "Generic (PLEG): container finished" podID="a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" containerID="4078ef8e9e930b90e52963fd33488744b48575ad03bc0edee8ee104b5cb41795" exitCode=0 Dec 06 10:59:16 crc kubenswrapper[4678]: I1206 10:59:16.651422 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9wvrw" event={"ID":"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46","Type":"ContainerDied","Data":"4078ef8e9e930b90e52963fd33488744b48575ad03bc0edee8ee104b5cb41795"} Dec 06 10:59:16 crc kubenswrapper[4678]: I1206 10:59:16.656212 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5a9c85b9-5f97-4449-ba76-0f81fd00cce1","Type":"ContainerStarted","Data":"b0774361ac765641b0ef673962afd75456f2a0f8f997247a57bfcd3156c2ab6c"} Dec 06 10:59:16 crc kubenswrapper[4678]: I1206 10:59:16.656421 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 06 10:59:16 crc kubenswrapper[4678]: I1206 10:59:16.715692 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.109191863 podStartE2EDuration="7.71566938s" podCreationTimestamp="2025-12-06 10:59:09 +0000 UTC" firstStartedPulling="2025-12-06 10:59:10.787803959 +0000 UTC m=+1355.631235398" lastFinishedPulling="2025-12-06 10:59:15.394281456 +0000 UTC m=+1360.237712915" observedRunningTime="2025-12-06 10:59:16.702613019 +0000 UTC m=+1361.546044478" watchObservedRunningTime="2025-12-06 10:59:16.71566938 +0000 UTC m=+1361.559100819" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.008376 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.136951 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-scripts\") pod \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.137189 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle\") pod \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.137256 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pshk6\" (UniqueName: \"kubernetes.io/projected/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-kube-api-access-pshk6\") pod \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.137286 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-config-data\") pod \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.142644 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-kube-api-access-pshk6" (OuterVolumeSpecName: "kube-api-access-pshk6") pod "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" (UID: "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46"). InnerVolumeSpecName "kube-api-access-pshk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.150719 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-scripts" (OuterVolumeSpecName: "scripts") pod "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" (UID: "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:18 crc kubenswrapper[4678]: E1206 10:59:18.172300 4678 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle podName:a26e1f26-4f6d-49c4-a66a-1ca2fc519d46 nodeName:}" failed. No retries permitted until 2025-12-06 10:59:18.671917018 +0000 UTC m=+1363.515348467 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle") pod "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" (UID: "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46") : error deleting /var/lib/kubelet/pods/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46/volume-subpaths: remove /var/lib/kubelet/pods/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46/volume-subpaths: no such file or directory Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.179627 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-config-data" (OuterVolumeSpecName: "config-data") pod "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" (UID: "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.239725 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pshk6\" (UniqueName: \"kubernetes.io/projected/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-kube-api-access-pshk6\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.239759 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.239770 4678 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-scripts\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.688177 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9wvrw" event={"ID":"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46","Type":"ContainerDied","Data":"9f90c0666324fd16bbaa1ff3fbafde5944780b42e3365d6c14cc05495cf90e73"} Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.688217 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f90c0666324fd16bbaa1ff3fbafde5944780b42e3365d6c14cc05495cf90e73" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.688286 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9wvrw" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.752072 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle\") pod \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\" (UID: \"a26e1f26-4f6d-49c4-a66a-1ca2fc519d46\") " Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.759574 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" (UID: "a26e1f26-4f6d-49c4-a66a-1ca2fc519d46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.855630 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.862907 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.863236 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-api" containerID="cri-o://435762cba1461d3cd78d31088be4ff9c7976bc87d2b1c212357b5b1d03f7bda0" gracePeriod=30 Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.863455 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-log" containerID="cri-o://4e4e6974c8af5150f499a772ee857cb6878a20594d12b388de39f7fc6c0be9df" gracePeriod=30 Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.888458 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.888763 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2c3a9d70-4738-4460-b779-5eebd9a06186" containerName="nova-scheduler-scheduler" containerID="cri-o://518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" gracePeriod=30 Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.962087 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.962665 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-log" containerID="cri-o://a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b" gracePeriod=30 Dec 06 10:59:18 crc kubenswrapper[4678]: I1206 10:59:18.963303 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-metadata" containerID="cri-o://014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67" gracePeriod=30 Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.699115 4678 generic.go:334] "Generic (PLEG): container finished" podID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerID="a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b" exitCode=143 Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.699165 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6fb98dd3-6574-44b5-a779-4c8c9a002637","Type":"ContainerDied","Data":"a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b"} Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.700927 4678 generic.go:334] "Generic (PLEG): container finished" podID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerID="435762cba1461d3cd78d31088be4ff9c7976bc87d2b1c212357b5b1d03f7bda0" exitCode=0 Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.700947 4678 generic.go:334] "Generic (PLEG): container finished" podID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerID="4e4e6974c8af5150f499a772ee857cb6878a20594d12b388de39f7fc6c0be9df" exitCode=143 Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.700962 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b79e75e4-bc70-46e9-b808-58a9a371e045","Type":"ContainerDied","Data":"435762cba1461d3cd78d31088be4ff9c7976bc87d2b1c212357b5b1d03f7bda0"} Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.700976 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b79e75e4-bc70-46e9-b808-58a9a371e045","Type":"ContainerDied","Data":"4e4e6974c8af5150f499a772ee857cb6878a20594d12b388de39f7fc6c0be9df"} Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.700985 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b79e75e4-bc70-46e9-b808-58a9a371e045","Type":"ContainerDied","Data":"e00fc95a020cc4a661fccbbdeee0dfe5a696f9a36984a5bb9ab9637e5c6a28ce"} Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.700995 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e00fc95a020cc4a661fccbbdeee0dfe5a696f9a36984a5bb9ab9637e5c6a28ce" Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.798123 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.976729 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79e75e4-bc70-46e9-b808-58a9a371e045-logs\") pod \"b79e75e4-bc70-46e9-b808-58a9a371e045\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.976790 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-combined-ca-bundle\") pod \"b79e75e4-bc70-46e9-b808-58a9a371e045\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.976823 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-public-tls-certs\") pod \"b79e75e4-bc70-46e9-b808-58a9a371e045\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.976969 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-internal-tls-certs\") pod \"b79e75e4-bc70-46e9-b808-58a9a371e045\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.977053 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-config-data\") pod \"b79e75e4-bc70-46e9-b808-58a9a371e045\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.977127 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9qdt\" (UniqueName: \"kubernetes.io/projected/b79e75e4-bc70-46e9-b808-58a9a371e045-kube-api-access-s9qdt\") pod \"b79e75e4-bc70-46e9-b808-58a9a371e045\" (UID: \"b79e75e4-bc70-46e9-b808-58a9a371e045\") " Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.977167 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b79e75e4-bc70-46e9-b808-58a9a371e045-logs" (OuterVolumeSpecName: "logs") pod "b79e75e4-bc70-46e9-b808-58a9a371e045" (UID: "b79e75e4-bc70-46e9-b808-58a9a371e045"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.977592 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b79e75e4-bc70-46e9-b808-58a9a371e045-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:19 crc kubenswrapper[4678]: I1206 10:59:19.983282 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b79e75e4-bc70-46e9-b808-58a9a371e045-kube-api-access-s9qdt" (OuterVolumeSpecName: "kube-api-access-s9qdt") pod "b79e75e4-bc70-46e9-b808-58a9a371e045" (UID: "b79e75e4-bc70-46e9-b808-58a9a371e045"). InnerVolumeSpecName "kube-api-access-s9qdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.018228 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b79e75e4-bc70-46e9-b808-58a9a371e045" (UID: "b79e75e4-bc70-46e9-b808-58a9a371e045"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.027011 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-config-data" (OuterVolumeSpecName: "config-data") pod "b79e75e4-bc70-46e9-b808-58a9a371e045" (UID: "b79e75e4-bc70-46e9-b808-58a9a371e045"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.052406 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b79e75e4-bc70-46e9-b808-58a9a371e045" (UID: "b79e75e4-bc70-46e9-b808-58a9a371e045"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.058754 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b79e75e4-bc70-46e9-b808-58a9a371e045" (UID: "b79e75e4-bc70-46e9-b808-58a9a371e045"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.079691 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9qdt\" (UniqueName: \"kubernetes.io/projected/b79e75e4-bc70-46e9-b808-58a9a371e045-kube-api-access-s9qdt\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.079740 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.079753 4678 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.079766 4678 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.079779 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b79e75e4-bc70-46e9-b808-58a9a371e045-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.707963 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.740295 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.750807 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.767901 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:20 crc kubenswrapper[4678]: E1206 10:59:20.768275 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69716076-6f0e-4432-91fc-82ae535dd96b" containerName="dnsmasq-dns" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768295 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="69716076-6f0e-4432-91fc-82ae535dd96b" containerName="dnsmasq-dns" Dec 06 10:59:20 crc kubenswrapper[4678]: E1206 10:59:20.768310 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" containerName="nova-manage" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768317 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" containerName="nova-manage" Dec 06 10:59:20 crc kubenswrapper[4678]: E1206 10:59:20.768330 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-log" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768336 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-log" Dec 06 10:59:20 crc kubenswrapper[4678]: E1206 10:59:20.768358 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-api" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768363 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-api" Dec 06 10:59:20 crc kubenswrapper[4678]: E1206 10:59:20.768376 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69716076-6f0e-4432-91fc-82ae535dd96b" containerName="init" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768382 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="69716076-6f0e-4432-91fc-82ae535dd96b" containerName="init" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768577 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="69716076-6f0e-4432-91fc-82ae535dd96b" containerName="dnsmasq-dns" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768595 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" containerName="nova-manage" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768608 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-log" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.768621 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" containerName="nova-api-api" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.769559 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.770993 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.771537 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.771735 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.777862 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.895056 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4859acf4-20ed-4af8-a83d-9c6303eec9d4-logs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.895155 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.895185 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtrvt\" (UniqueName: \"kubernetes.io/projected/4859acf4-20ed-4af8-a83d-9c6303eec9d4-kube-api-access-qtrvt\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.895236 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-public-tls-certs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.895256 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.895282 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-config-data\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.997430 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-config-data\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.997552 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4859acf4-20ed-4af8-a83d-9c6303eec9d4-logs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.997637 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.997669 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtrvt\" (UniqueName: \"kubernetes.io/projected/4859acf4-20ed-4af8-a83d-9c6303eec9d4-kube-api-access-qtrvt\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.997719 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-public-tls-certs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.997738 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:20 crc kubenswrapper[4678]: I1206 10:59:20.998102 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4859acf4-20ed-4af8-a83d-9c6303eec9d4-logs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.002136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-public-tls-certs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.008076 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-config-data\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.008629 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.018819 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4859acf4-20ed-4af8-a83d-9c6303eec9d4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.032284 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtrvt\" (UniqueName: \"kubernetes.io/projected/4859acf4-20ed-4af8-a83d-9c6303eec9d4-kube-api-access-qtrvt\") pod \"nova-api-0\" (UID: \"4859acf4-20ed-4af8-a83d-9c6303eec9d4\") " pod="openstack/nova-api-0" Dec 06 10:59:21 crc kubenswrapper[4678]: E1206 10:59:21.060640 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 10:59:21 crc kubenswrapper[4678]: E1206 10:59:21.063792 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 10:59:21 crc kubenswrapper[4678]: E1206 10:59:21.065409 4678 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 06 10:59:21 crc kubenswrapper[4678]: E1206 10:59:21.065495 4678 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2c3a9d70-4738-4460-b779-5eebd9a06186" containerName="nova-scheduler-scheduler" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.089872 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.491287 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b79e75e4-bc70-46e9-b808-58a9a371e045" path="/var/lib/kubelet/pods/b79e75e4-bc70-46e9-b808-58a9a371e045/volumes" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.632908 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.644230 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.741621 4678 generic.go:334] "Generic (PLEG): container finished" podID="2c3a9d70-4738-4460-b779-5eebd9a06186" containerID="518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" exitCode=0 Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.741737 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c3a9d70-4738-4460-b779-5eebd9a06186","Type":"ContainerDied","Data":"518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e"} Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.741751 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.741779 4678 scope.go:117] "RemoveContainer" containerID="518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.741766 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2c3a9d70-4738-4460-b779-5eebd9a06186","Type":"ContainerDied","Data":"ba15f794fe9537757ddb1cf79bb6f7a7d7e332c8bf769e505f851a66e64c1529"} Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.745008 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4859acf4-20ed-4af8-a83d-9c6303eec9d4","Type":"ContainerStarted","Data":"8643decb03ce2c55362cbe114e7bbe77bf448ece09be7ceb3b4615ce03defe9f"} Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.769303 4678 scope.go:117] "RemoveContainer" containerID="518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" Dec 06 10:59:21 crc kubenswrapper[4678]: E1206 10:59:21.771665 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e\": container with ID starting with 518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e not found: ID does not exist" containerID="518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.771853 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e"} err="failed to get container status \"518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e\": rpc error: code = NotFound desc = could not find container \"518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e\": container with ID starting with 518f576212f04c309b5fd23e3f18254631dde7d22a2175c5bea3a0f56ee0c97e not found: ID does not exist" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.812657 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbfsb\" (UniqueName: \"kubernetes.io/projected/2c3a9d70-4738-4460-b779-5eebd9a06186-kube-api-access-kbfsb\") pod \"2c3a9d70-4738-4460-b779-5eebd9a06186\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.813440 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-combined-ca-bundle\") pod \"2c3a9d70-4738-4460-b779-5eebd9a06186\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.816822 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-config-data\") pod \"2c3a9d70-4738-4460-b779-5eebd9a06186\" (UID: \"2c3a9d70-4738-4460-b779-5eebd9a06186\") " Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.818286 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c3a9d70-4738-4460-b779-5eebd9a06186-kube-api-access-kbfsb" (OuterVolumeSpecName: "kube-api-access-kbfsb") pod "2c3a9d70-4738-4460-b779-5eebd9a06186" (UID: "2c3a9d70-4738-4460-b779-5eebd9a06186"). InnerVolumeSpecName "kube-api-access-kbfsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.856284 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2c3a9d70-4738-4460-b779-5eebd9a06186" (UID: "2c3a9d70-4738-4460-b779-5eebd9a06186"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.864869 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-config-data" (OuterVolumeSpecName: "config-data") pod "2c3a9d70-4738-4460-b779-5eebd9a06186" (UID: "2c3a9d70-4738-4460-b779-5eebd9a06186"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.920460 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbfsb\" (UniqueName: \"kubernetes.io/projected/2c3a9d70-4738-4460-b779-5eebd9a06186-kube-api-access-kbfsb\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.920522 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:21 crc kubenswrapper[4678]: I1206 10:59:21.920534 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c3a9d70-4738-4460-b779-5eebd9a06186-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.087375 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.161652 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.198863 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:59:22 crc kubenswrapper[4678]: E1206 10:59:22.199310 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c3a9d70-4738-4460-b779-5eebd9a06186" containerName="nova-scheduler-scheduler" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.199329 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c3a9d70-4738-4460-b779-5eebd9a06186" containerName="nova-scheduler-scheduler" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.199562 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c3a9d70-4738-4460-b779-5eebd9a06186" containerName="nova-scheduler-scheduler" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.200243 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.203983 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.208653 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.330859 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-778nl\" (UniqueName: \"kubernetes.io/projected/a2fb742d-13dd-468a-918e-37c53f81b294-kube-api-access-778nl\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.331056 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fb742d-13dd-468a-918e-37c53f81b294-config-data\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.331121 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fb742d-13dd-468a-918e-37c53f81b294-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: E1206 10:59:22.416100 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fb98dd3_6574_44b5_a779_4c8c9a002637.slice/crio-conmon-014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fb98dd3_6574_44b5_a779_4c8c9a002637.slice/crio-014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67.scope\": RecentStats: unable to find data in memory cache]" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.433713 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fb742d-13dd-468a-918e-37c53f81b294-config-data\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.433780 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fb742d-13dd-468a-918e-37c53f81b294-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.433854 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-778nl\" (UniqueName: \"kubernetes.io/projected/a2fb742d-13dd-468a-918e-37c53f81b294-kube-api-access-778nl\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.492160 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2fb742d-13dd-468a-918e-37c53f81b294-config-data\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.492568 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2fb742d-13dd-468a-918e-37c53f81b294-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.492836 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-778nl\" (UniqueName: \"kubernetes.io/projected/a2fb742d-13dd-468a-918e-37c53f81b294-kube-api-access-778nl\") pod \"nova-scheduler-0\" (UID: \"a2fb742d-13dd-468a-918e-37c53f81b294\") " pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.614057 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.628851 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.739383 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-nova-metadata-tls-certs\") pod \"6fb98dd3-6574-44b5-a779-4c8c9a002637\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.739850 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-config-data\") pod \"6fb98dd3-6574-44b5-a779-4c8c9a002637\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.739893 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fb98dd3-6574-44b5-a779-4c8c9a002637-logs\") pod \"6fb98dd3-6574-44b5-a779-4c8c9a002637\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.739967 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-combined-ca-bundle\") pod \"6fb98dd3-6574-44b5-a779-4c8c9a002637\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.740018 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttzkb\" (UniqueName: \"kubernetes.io/projected/6fb98dd3-6574-44b5-a779-4c8c9a002637-kube-api-access-ttzkb\") pod \"6fb98dd3-6574-44b5-a779-4c8c9a002637\" (UID: \"6fb98dd3-6574-44b5-a779-4c8c9a002637\") " Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.743080 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fb98dd3-6574-44b5-a779-4c8c9a002637-logs" (OuterVolumeSpecName: "logs") pod "6fb98dd3-6574-44b5-a779-4c8c9a002637" (UID: "6fb98dd3-6574-44b5-a779-4c8c9a002637"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.743193 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fb98dd3-6574-44b5-a779-4c8c9a002637-kube-api-access-ttzkb" (OuterVolumeSpecName: "kube-api-access-ttzkb") pod "6fb98dd3-6574-44b5-a779-4c8c9a002637" (UID: "6fb98dd3-6574-44b5-a779-4c8c9a002637"). InnerVolumeSpecName "kube-api-access-ttzkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.773947 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4859acf4-20ed-4af8-a83d-9c6303eec9d4","Type":"ContainerStarted","Data":"25f25ccfc433ed3015b70ef8f52358d0cc696982cb569cedc2ff48e2f8af3017"} Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.773985 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4859acf4-20ed-4af8-a83d-9c6303eec9d4","Type":"ContainerStarted","Data":"7caaa811deee02404980e24e76e1637a8fea73ceff594092499b370f310a5968"} Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.779269 4678 generic.go:334] "Generic (PLEG): container finished" podID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerID="014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67" exitCode=0 Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.779330 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6fb98dd3-6574-44b5-a779-4c8c9a002637","Type":"ContainerDied","Data":"014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67"} Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.779355 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6fb98dd3-6574-44b5-a779-4c8c9a002637","Type":"ContainerDied","Data":"1d328e629e6e6e419fe44ad8333365d831322e75c4a619a7a2e0630c3a865087"} Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.779370 4678 scope.go:117] "RemoveContainer" containerID="014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.779500 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.809946 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-config-data" (OuterVolumeSpecName: "config-data") pod "6fb98dd3-6574-44b5-a779-4c8c9a002637" (UID: "6fb98dd3-6574-44b5-a779-4c8c9a002637"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.821103 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fb98dd3-6574-44b5-a779-4c8c9a002637" (UID: "6fb98dd3-6574-44b5-a779-4c8c9a002637"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.825090 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.825068983 podStartE2EDuration="2.825068983s" podCreationTimestamp="2025-12-06 10:59:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:59:22.801662502 +0000 UTC m=+1367.645093931" watchObservedRunningTime="2025-12-06 10:59:22.825068983 +0000 UTC m=+1367.668500422" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.837214 4678 scope.go:117] "RemoveContainer" containerID="a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.842092 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.842120 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttzkb\" (UniqueName: \"kubernetes.io/projected/6fb98dd3-6574-44b5-a779-4c8c9a002637-kube-api-access-ttzkb\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.842135 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.842145 4678 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fb98dd3-6574-44b5-a779-4c8c9a002637-logs\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.863833 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6fb98dd3-6574-44b5-a779-4c8c9a002637" (UID: "6fb98dd3-6574-44b5-a779-4c8c9a002637"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.871693 4678 scope.go:117] "RemoveContainer" containerID="014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67" Dec 06 10:59:22 crc kubenswrapper[4678]: E1206 10:59:22.872393 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67\": container with ID starting with 014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67 not found: ID does not exist" containerID="014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.872430 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67"} err="failed to get container status \"014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67\": rpc error: code = NotFound desc = could not find container \"014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67\": container with ID starting with 014df4eff063f289c71bdc3f9c73bc161e2ec5e51211a9906833aee1f2ce5c67 not found: ID does not exist" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.872456 4678 scope.go:117] "RemoveContainer" containerID="a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b" Dec 06 10:59:22 crc kubenswrapper[4678]: E1206 10:59:22.874852 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b\": container with ID starting with a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b not found: ID does not exist" containerID="a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.874892 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b"} err="failed to get container status \"a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b\": rpc error: code = NotFound desc = could not find container \"a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b\": container with ID starting with a99b0fd78bf0cb4844002ee79c2675dd8b602aaca527b93ea00e59add37aff2b not found: ID does not exist" Dec 06 10:59:22 crc kubenswrapper[4678]: I1206 10:59:22.944269 4678 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6fb98dd3-6574-44b5-a779-4c8c9a002637-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.132672 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.147882 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.187258 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.207076 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:59:23 crc kubenswrapper[4678]: E1206 10:59:23.207739 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-log" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.207761 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-log" Dec 06 10:59:23 crc kubenswrapper[4678]: E1206 10:59:23.207803 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-metadata" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.207812 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-metadata" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.208033 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-metadata" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.208068 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" containerName="nova-metadata-log" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.209154 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.210600 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.211819 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.235562 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.351227 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63c277fe-de03-4243-98a2-8d7efaa92c4a-logs\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.351369 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-config-data\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.351442 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65l9c\" (UniqueName: \"kubernetes.io/projected/63c277fe-de03-4243-98a2-8d7efaa92c4a-kube-api-access-65l9c\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.351558 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.351753 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.453893 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-config-data\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.453949 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65l9c\" (UniqueName: \"kubernetes.io/projected/63c277fe-de03-4243-98a2-8d7efaa92c4a-kube-api-access-65l9c\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.453979 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.454028 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.454100 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63c277fe-de03-4243-98a2-8d7efaa92c4a-logs\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.454659 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/63c277fe-de03-4243-98a2-8d7efaa92c4a-logs\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.457534 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.457632 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.465298 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/63c277fe-de03-4243-98a2-8d7efaa92c4a-config-data\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.486807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65l9c\" (UniqueName: \"kubernetes.io/projected/63c277fe-de03-4243-98a2-8d7efaa92c4a-kube-api-access-65l9c\") pod \"nova-metadata-0\" (UID: \"63c277fe-de03-4243-98a2-8d7efaa92c4a\") " pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.490983 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c3a9d70-4738-4460-b779-5eebd9a06186" path="/var/lib/kubelet/pods/2c3a9d70-4738-4460-b779-5eebd9a06186/volumes" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.491757 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fb98dd3-6574-44b5-a779-4c8c9a002637" path="/var/lib/kubelet/pods/6fb98dd3-6574-44b5-a779-4c8c9a002637/volumes" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.634878 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.832759 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a2fb742d-13dd-468a-918e-37c53f81b294","Type":"ContainerStarted","Data":"6bcd34e000f1e6357923bc71b27a244fe37d378dcb120ccc485a07db7747d2d6"} Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.832853 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a2fb742d-13dd-468a-918e-37c53f81b294","Type":"ContainerStarted","Data":"0499d52086b78313d714ac17a4f2c8bf32fe4fecf043b8bedbf437fd321a9d80"} Dec 06 10:59:23 crc kubenswrapper[4678]: I1206 10:59:23.863438 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.863411668 podStartE2EDuration="1.863411668s" podCreationTimestamp="2025-12-06 10:59:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:59:23.86314592 +0000 UTC m=+1368.706577369" watchObservedRunningTime="2025-12-06 10:59:23.863411668 +0000 UTC m=+1368.706843117" Dec 06 10:59:24 crc kubenswrapper[4678]: I1206 10:59:24.184000 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 06 10:59:24 crc kubenswrapper[4678]: I1206 10:59:24.841577 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63c277fe-de03-4243-98a2-8d7efaa92c4a","Type":"ContainerStarted","Data":"7aa1862826ff137636eb8d544bc6f725e8c8e162b0ee2194506bd6b57e32ffb1"} Dec 06 10:59:24 crc kubenswrapper[4678]: I1206 10:59:24.841861 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63c277fe-de03-4243-98a2-8d7efaa92c4a","Type":"ContainerStarted","Data":"60d793647742e1bbb2f806505cfed60e117e02c38f15ddfc4864c3d33d6775f5"} Dec 06 10:59:24 crc kubenswrapper[4678]: I1206 10:59:24.841878 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"63c277fe-de03-4243-98a2-8d7efaa92c4a","Type":"ContainerStarted","Data":"1085d8dd03816e0dbc257d9222518db428596ad5c0290f544336bd69260124f9"} Dec 06 10:59:24 crc kubenswrapper[4678]: I1206 10:59:24.870804 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.870779997 podStartE2EDuration="1.870779997s" podCreationTimestamp="2025-12-06 10:59:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 10:59:24.860980199 +0000 UTC m=+1369.704411648" watchObservedRunningTime="2025-12-06 10:59:24.870779997 +0000 UTC m=+1369.714211436" Dec 06 10:59:27 crc kubenswrapper[4678]: I1206 10:59:27.629537 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 06 10:59:28 crc kubenswrapper[4678]: I1206 10:59:28.635431 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 10:59:28 crc kubenswrapper[4678]: I1206 10:59:28.636464 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.505478 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.505747 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.505856 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.506622 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c7aaeb71fc254443c399a6d76a377d0a6c35a168b9037ac2247e9eecdf49d005"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.506802 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://c7aaeb71fc254443c399a6d76a377d0a6c35a168b9037ac2247e9eecdf49d005" gracePeriod=600 Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.887918 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="c7aaeb71fc254443c399a6d76a377d0a6c35a168b9037ac2247e9eecdf49d005" exitCode=0 Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.887968 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"c7aaeb71fc254443c399a6d76a377d0a6c35a168b9037ac2247e9eecdf49d005"} Dec 06 10:59:29 crc kubenswrapper[4678]: I1206 10:59:29.888003 4678 scope.go:117] "RemoveContainer" containerID="996baf59448dd4fd88041c7dc359f0a860305f2ace3dab71db17f1d751ccc5a8" Dec 06 10:59:30 crc kubenswrapper[4678]: I1206 10:59:30.900784 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da"} Dec 06 10:59:31 crc kubenswrapper[4678]: I1206 10:59:31.091228 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 10:59:31 crc kubenswrapper[4678]: I1206 10:59:31.091844 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 06 10:59:32 crc kubenswrapper[4678]: I1206 10:59:32.105768 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4859acf4-20ed-4af8-a83d-9c6303eec9d4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:59:32 crc kubenswrapper[4678]: I1206 10:59:32.105816 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4859acf4-20ed-4af8-a83d-9c6303eec9d4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:59:32 crc kubenswrapper[4678]: I1206 10:59:32.629271 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 06 10:59:32 crc kubenswrapper[4678]: I1206 10:59:32.669143 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 06 10:59:32 crc kubenswrapper[4678]: I1206 10:59:32.945081 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 06 10:59:33 crc kubenswrapper[4678]: I1206 10:59:33.635558 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 10:59:33 crc kubenswrapper[4678]: I1206 10:59:33.635721 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 06 10:59:34 crc kubenswrapper[4678]: I1206 10:59:34.682917 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="63c277fe-de03-4243-98a2-8d7efaa92c4a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:59:34 crc kubenswrapper[4678]: I1206 10:59:34.683640 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="63c277fe-de03-4243-98a2-8d7efaa92c4a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 06 10:59:39 crc kubenswrapper[4678]: I1206 10:59:39.929619 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 06 10:59:41 crc kubenswrapper[4678]: I1206 10:59:41.105757 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 10:59:41 crc kubenswrapper[4678]: I1206 10:59:41.106603 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 10:59:41 crc kubenswrapper[4678]: I1206 10:59:41.115746 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 06 10:59:41 crc kubenswrapper[4678]: I1206 10:59:41.128577 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 10:59:42 crc kubenswrapper[4678]: I1206 10:59:42.013422 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 06 10:59:42 crc kubenswrapper[4678]: I1206 10:59:42.025643 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 06 10:59:43 crc kubenswrapper[4678]: I1206 10:59:43.642301 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 10:59:43 crc kubenswrapper[4678]: I1206 10:59:43.642729 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 06 10:59:43 crc kubenswrapper[4678]: I1206 10:59:43.648022 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 10:59:43 crc kubenswrapper[4678]: I1206 10:59:43.652437 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 06 10:59:51 crc kubenswrapper[4678]: I1206 10:59:51.601175 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 10:59:52 crc kubenswrapper[4678]: I1206 10:59:52.453364 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 10:59:56 crc kubenswrapper[4678]: I1206 10:59:56.430114 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerName="rabbitmq" containerID="cri-o://02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4" gracePeriod=604796 Dec 06 10:59:56 crc kubenswrapper[4678]: I1206 10:59:56.999079 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="995bfee3-8461-4774-aa52-576dc0eacbda" containerName="rabbitmq" containerID="cri-o://649f042958a1f97564fc9864785c913d1ae04d99007549421e0d8f0b8e4378ef" gracePeriod=604796 Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.164952 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2"] Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.166917 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.170802 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.173052 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.176728 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2"] Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.336894 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/096244e2-6040-4775-a2c3-29da87ed9817-config-volume\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.337176 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/096244e2-6040-4775-a2c3-29da87ed9817-secret-volume\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.337247 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p564\" (UniqueName: \"kubernetes.io/projected/096244e2-6040-4775-a2c3-29da87ed9817-kube-api-access-9p564\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.438609 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/096244e2-6040-4775-a2c3-29da87ed9817-secret-volume\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.438701 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p564\" (UniqueName: \"kubernetes.io/projected/096244e2-6040-4775-a2c3-29da87ed9817-kube-api-access-9p564\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.438831 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/096244e2-6040-4775-a2c3-29da87ed9817-config-volume\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.440036 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/096244e2-6040-4775-a2c3-29da87ed9817-config-volume\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.444943 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/096244e2-6040-4775-a2c3-29da87ed9817-secret-volume\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.455787 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p564\" (UniqueName: \"kubernetes.io/projected/096244e2-6040-4775-a2c3-29da87ed9817-kube-api-access-9p564\") pod \"collect-profiles-29416980-x7ft2\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:00 crc kubenswrapper[4678]: I1206 11:00:00.513222 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:01 crc kubenswrapper[4678]: I1206 11:00:01.017539 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2"] Dec 06 11:00:01 crc kubenswrapper[4678]: I1206 11:00:01.217073 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" event={"ID":"096244e2-6040-4775-a2c3-29da87ed9817","Type":"ContainerStarted","Data":"b8a5cad4925949cd33b025ecefb6d9e5e2ea129563692cf6c0f2269fd62fc4b9"} Dec 06 11:00:01 crc kubenswrapper[4678]: I1206 11:00:01.217119 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" event={"ID":"096244e2-6040-4775-a2c3-29da87ed9817","Type":"ContainerStarted","Data":"ba8e4ae0e3366b6d916aae8b86b0def203e35daa7ab9a77a83cbb412ee896a10"} Dec 06 11:00:02 crc kubenswrapper[4678]: I1206 11:00:02.238185 4678 generic.go:334] "Generic (PLEG): container finished" podID="096244e2-6040-4775-a2c3-29da87ed9817" containerID="b8a5cad4925949cd33b025ecefb6d9e5e2ea129563692cf6c0f2269fd62fc4b9" exitCode=0 Dec 06 11:00:02 crc kubenswrapper[4678]: I1206 11:00:02.238243 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" event={"ID":"096244e2-6040-4775-a2c3-29da87ed9817","Type":"ContainerDied","Data":"b8a5cad4925949cd33b025ecefb6d9e5e2ea129563692cf6c0f2269fd62fc4b9"} Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.116999 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.272207 4678 generic.go:334] "Generic (PLEG): container finished" podID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerID="02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4" exitCode=0 Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.272272 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52efc5a4-585a-4b9e-93f9-f40527e9a6c9","Type":"ContainerDied","Data":"02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4"} Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.272301 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"52efc5a4-585a-4b9e-93f9-f40527e9a6c9","Type":"ContainerDied","Data":"418dac276e67317edebb84bd03d88223120391184b5c2bad6b3dafd2eea07018"} Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.272329 4678 scope.go:117] "RemoveContainer" containerID="02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.272329 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.283548 4678 generic.go:334] "Generic (PLEG): container finished" podID="995bfee3-8461-4774-aa52-576dc0eacbda" containerID="649f042958a1f97564fc9864785c913d1ae04d99007549421e0d8f0b8e4378ef" exitCode=0 Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.283875 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"995bfee3-8461-4774-aa52-576dc0eacbda","Type":"ContainerDied","Data":"649f042958a1f97564fc9864785c913d1ae04d99007549421e0d8f0b8e4378ef"} Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294143 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294226 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-pod-info\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294263 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-config-data\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294301 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-tls\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294332 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-confd\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294350 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-erlang-cookie-secret\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294423 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-erlang-cookie\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294480 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-plugins-conf\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294522 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhk62\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-kube-api-access-rhk62\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294541 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-server-conf\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.294560 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-plugins\") pod \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\" (UID: \"52efc5a4-585a-4b9e-93f9-f40527e9a6c9\") " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.296728 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.298715 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.307296 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.314668 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.328573 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.337748 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-kube-api-access-rhk62" (OuterVolumeSpecName: "kube-api-access-rhk62") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "kube-api-access-rhk62". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.392746 4678 scope.go:117] "RemoveContainer" containerID="f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.396321 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.396658 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.396746 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.396821 4678 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.396878 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhk62\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-kube-api-access-rhk62\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.396947 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.412340 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-pod-info" (OuterVolumeSpecName: "pod-info") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.413323 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.423246 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.481633 4678 scope.go:117] "RemoveContainer" containerID="02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4" Dec 06 11:00:03 crc kubenswrapper[4678]: E1206 11:00:03.485336 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4\": container with ID starting with 02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4 not found: ID does not exist" containerID="02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.485393 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4"} err="failed to get container status \"02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4\": rpc error: code = NotFound desc = could not find container \"02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4\": container with ID starting with 02f364aa9b5665b029e8ab1115e94f2fa5b5cff835be42ec41de72face18a0f4 not found: ID does not exist" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.485421 4678 scope.go:117] "RemoveContainer" containerID="f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b" Dec 06 11:00:03 crc kubenswrapper[4678]: E1206 11:00:03.487698 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b\": container with ID starting with f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b not found: ID does not exist" containerID="f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.487748 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b"} err="failed to get container status \"f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b\": rpc error: code = NotFound desc = could not find container \"f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b\": container with ID starting with f74294b55db84320ee6034b0c378cb5aca7d35d98ad37399043b400819f1f91b not found: ID does not exist" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.490150 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-server-conf" (OuterVolumeSpecName: "server-conf") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.490317 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-config-data" (OuterVolumeSpecName: "config-data") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.498777 4678 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.498814 4678 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.498827 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.498837 4678 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.498850 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.564142 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "52efc5a4-585a-4b9e-93f9-f40527e9a6c9" (UID: "52efc5a4-585a-4b9e-93f9-f40527e9a6c9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.594207 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-4zcbf"] Dec 06 11:00:03 crc kubenswrapper[4678]: E1206 11:00:03.594602 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerName="rabbitmq" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.594620 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerName="rabbitmq" Dec 06 11:00:03 crc kubenswrapper[4678]: E1206 11:00:03.594657 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerName="setup-container" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.594665 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerName="setup-container" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.594836 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" containerName="rabbitmq" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.595727 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.605601 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.605813 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/52efc5a4-585a-4b9e-93f9-f40527e9a6c9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.629904 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-4zcbf"] Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.708580 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-svc\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.715356 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p42qt\" (UniqueName: \"kubernetes.io/projected/4dac7d51-741a-418d-8eb4-e4583fe9cc63-kube-api-access-p42qt\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.715514 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.715968 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-config\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.716208 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.716497 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.716625 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: E1206 11:00:03.765422 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52efc5a4_585a_4b9e_93f9_f40527e9a6c9.slice/crio-418dac276e67317edebb84bd03d88223120391184b5c2bad6b3dafd2eea07018\": RecentStats: unable to find data in memory cache]" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.787635 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.817856 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.818195 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.818261 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-svc\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.818280 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p42qt\" (UniqueName: \"kubernetes.io/projected/4dac7d51-741a-418d-8eb4-e4583fe9cc63-kube-api-access-p42qt\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.818295 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.818369 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-config\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.818391 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.819138 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.819662 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.820153 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.820657 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-svc\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.821369 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-config\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.822476 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.841457 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.854300 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.854959 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p42qt\" (UniqueName: \"kubernetes.io/projected/4dac7d51-741a-418d-8eb4-e4583fe9cc63-kube-api-access-p42qt\") pod \"dnsmasq-dns-d558885bc-4zcbf\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.858074 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.862795 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.863057 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.863169 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.863189 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zmxm7" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.863223 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.863649 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.863778 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 06 11:00:03 crc kubenswrapper[4678]: I1206 11:00:03.887119 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.014869 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.018809 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037518 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037588 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037641 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037665 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037710 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv2x9\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-kube-api-access-vv2x9\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037742 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037775 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037802 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037827 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-config-data\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037865 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.037914 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.116211 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.138756 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/096244e2-6040-4775-a2c3-29da87ed9817-secret-volume\") pod \"096244e2-6040-4775-a2c3-29da87ed9817\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139272 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9p564\" (UniqueName: \"kubernetes.io/projected/096244e2-6040-4775-a2c3-29da87ed9817-kube-api-access-9p564\") pod \"096244e2-6040-4775-a2c3-29da87ed9817\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139327 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/096244e2-6040-4775-a2c3-29da87ed9817-config-volume\") pod \"096244e2-6040-4775-a2c3-29da87ed9817\" (UID: \"096244e2-6040-4775-a2c3-29da87ed9817\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139696 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139732 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139760 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139782 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139822 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv2x9\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-kube-api-access-vv2x9\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139852 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139880 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139902 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139923 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-config-data\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139955 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.139997 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.141275 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.153799 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.163183 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.168187 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.171362 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/096244e2-6040-4775-a2c3-29da87ed9817-config-volume" (OuterVolumeSpecName: "config-volume") pod "096244e2-6040-4775-a2c3-29da87ed9817" (UID: "096244e2-6040-4775-a2c3-29da87ed9817"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.171737 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.171857 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.172512 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-config-data\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.182763 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.185910 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.191010 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.197638 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/096244e2-6040-4775-a2c3-29da87ed9817-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "096244e2-6040-4775-a2c3-29da87ed9817" (UID: "096244e2-6040-4775-a2c3-29da87ed9817"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.198004 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv2x9\" (UniqueName: \"kubernetes.io/projected/cb4272e8-bb6c-43d4-8539-baeb4f3c14a8-kube-api-access-vv2x9\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.200706 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096244e2-6040-4775-a2c3-29da87ed9817-kube-api-access-9p564" (OuterVolumeSpecName: "kube-api-access-9p564") pod "096244e2-6040-4775-a2c3-29da87ed9817" (UID: "096244e2-6040-4775-a2c3-29da87ed9817"). InnerVolumeSpecName "kube-api-access-9p564". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.240849 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/995bfee3-8461-4774-aa52-576dc0eacbda-erlang-cookie-secret\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.240910 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.240985 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-config-data\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241052 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-plugins-conf\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241118 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-tls\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241181 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksdhv\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-kube-api-access-ksdhv\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241221 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-erlang-cookie\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241255 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/995bfee3-8461-4774-aa52-576dc0eacbda-pod-info\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241278 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-confd\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241345 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-server-conf\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.241414 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-plugins\") pod \"995bfee3-8461-4774-aa52-576dc0eacbda\" (UID: \"995bfee3-8461-4774-aa52-576dc0eacbda\") " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.245380 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.245992 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-kube-api-access-ksdhv" (OuterVolumeSpecName: "kube-api-access-ksdhv") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "kube-api-access-ksdhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.248812 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksdhv\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-kube-api-access-ksdhv\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.248850 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9p564\" (UniqueName: \"kubernetes.io/projected/096244e2-6040-4775-a2c3-29da87ed9817-kube-api-access-9p564\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.248864 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/096244e2-6040-4775-a2c3-29da87ed9817-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.248876 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/096244e2-6040-4775-a2c3-29da87ed9817-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.248891 4678 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.249022 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.251000 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.252945 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.253527 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/995bfee3-8461-4774-aa52-576dc0eacbda-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.257603 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.276561 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/995bfee3-8461-4774-aa52-576dc0eacbda-pod-info" (OuterVolumeSpecName: "pod-info") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.289136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8\") " pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.312208 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"995bfee3-8461-4774-aa52-576dc0eacbda","Type":"ContainerDied","Data":"80a760531885b8a27aec23e2d66042f4ee95d2b60da078527e4702a040e43caa"} Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.312452 4678 scope.go:117] "RemoveContainer" containerID="649f042958a1f97564fc9864785c913d1ae04d99007549421e0d8f0b8e4378ef" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.313095 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.314395 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.316560 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-config-data" (OuterVolumeSpecName: "config-data") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.334208 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" event={"ID":"096244e2-6040-4775-a2c3-29da87ed9817","Type":"ContainerDied","Data":"ba8e4ae0e3366b6d916aae8b86b0def203e35daa7ab9a77a83cbb412ee896a10"} Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.334429 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba8e4ae0e3366b6d916aae8b86b0def203e35daa7ab9a77a83cbb412ee896a10" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.334300 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.359090 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.359117 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.359128 4678 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/995bfee3-8461-4774-aa52-576dc0eacbda-pod-info\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.359140 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.359148 4678 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/995bfee3-8461-4774-aa52-576dc0eacbda-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.359169 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.359178 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.362782 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-server-conf" (OuterVolumeSpecName: "server-conf") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.375214 4678 scope.go:117] "RemoveContainer" containerID="a67896673bf325c989b177f147cde987c7fc2280ee8ac8bf78a9c554bbb3c2b7" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.431996 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.460777 4678 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/995bfee3-8461-4774-aa52-576dc0eacbda-server-conf\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.460815 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.513699 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "995bfee3-8461-4774-aa52-576dc0eacbda" (UID: "995bfee3-8461-4774-aa52-576dc0eacbda"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.562712 4678 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/995bfee3-8461-4774-aa52-576dc0eacbda-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.666549 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-4zcbf"] Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.713478 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.723082 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.777054 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 11:00:04 crc kubenswrapper[4678]: E1206 11:00:04.777426 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995bfee3-8461-4774-aa52-576dc0eacbda" containerName="rabbitmq" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.777439 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="995bfee3-8461-4774-aa52-576dc0eacbda" containerName="rabbitmq" Dec 06 11:00:04 crc kubenswrapper[4678]: E1206 11:00:04.777452 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096244e2-6040-4775-a2c3-29da87ed9817" containerName="collect-profiles" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.777460 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="096244e2-6040-4775-a2c3-29da87ed9817" containerName="collect-profiles" Dec 06 11:00:04 crc kubenswrapper[4678]: E1206 11:00:04.777513 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="995bfee3-8461-4774-aa52-576dc0eacbda" containerName="setup-container" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.777520 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="995bfee3-8461-4774-aa52-576dc0eacbda" containerName="setup-container" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.777690 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="096244e2-6040-4775-a2c3-29da87ed9817" containerName="collect-profiles" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.777701 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="995bfee3-8461-4774-aa52-576dc0eacbda" containerName="rabbitmq" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.778623 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.819633 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.819821 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-p6xng" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.819969 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.820184 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.824148 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.824650 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.824766 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.872769 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.872836 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.872858 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8jmr\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-kube-api-access-g8jmr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.872950 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.872983 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.873039 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.873095 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.879906 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.879968 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.879994 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.880019 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:04 crc kubenswrapper[4678]: I1206 11:00:04.921603 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002098 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002144 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002190 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002234 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002257 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002275 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002292 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002312 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002390 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002415 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.002548 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8jmr\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-kube-api-access-g8jmr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.007898 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.010601 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.013376 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.017992 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.022641 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.022828 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.024197 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.024725 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.025627 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.035648 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.040939 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8jmr\" (UniqueName: \"kubernetes.io/projected/f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3-kube-api-access-g8jmr\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.122596 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3\") " pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.331635 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.359362 4678 generic.go:334] "Generic (PLEG): container finished" podID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerID="534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d" exitCode=0 Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.359419 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" event={"ID":"4dac7d51-741a-418d-8eb4-e4583fe9cc63","Type":"ContainerDied","Data":"534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d"} Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.359447 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" event={"ID":"4dac7d51-741a-418d-8eb4-e4583fe9cc63","Type":"ContainerStarted","Data":"0770d47c7b9dbe492e9dc9b4024262cfc4f084d7b99a2eab00ac9cf185668e66"} Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.437620 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.563989 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52efc5a4-585a-4b9e-93f9-f40527e9a6c9" path="/var/lib/kubelet/pods/52efc5a4-585a-4b9e-93f9-f40527e9a6c9/volumes" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.571144 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="995bfee3-8461-4774-aa52-576dc0eacbda" path="/var/lib/kubelet/pods/995bfee3-8461-4774-aa52-576dc0eacbda/volumes" Dec 06 11:00:05 crc kubenswrapper[4678]: I1206 11:00:05.880793 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 06 11:00:05 crc kubenswrapper[4678]: W1206 11:00:05.885993 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7fe7aa1_62b5_4c6b_be2e_b64fce2711b3.slice/crio-eccfac74d34ed634fe037e79cb9315bc5b30df9b0f731e290b3b87f9af9b75a9 WatchSource:0}: Error finding container eccfac74d34ed634fe037e79cb9315bc5b30df9b0f731e290b3b87f9af9b75a9: Status 404 returned error can't find the container with id eccfac74d34ed634fe037e79cb9315bc5b30df9b0f731e290b3b87f9af9b75a9 Dec 06 11:00:06 crc kubenswrapper[4678]: I1206 11:00:06.371124 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8","Type":"ContainerStarted","Data":"f004bf92e3d85e371441d53d363fac2533ebe65693b5f51aaf8772e268a300ad"} Dec 06 11:00:06 crc kubenswrapper[4678]: I1206 11:00:06.379586 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" event={"ID":"4dac7d51-741a-418d-8eb4-e4583fe9cc63","Type":"ContainerStarted","Data":"3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199"} Dec 06 11:00:06 crc kubenswrapper[4678]: I1206 11:00:06.379651 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:06 crc kubenswrapper[4678]: I1206 11:00:06.381312 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3","Type":"ContainerStarted","Data":"eccfac74d34ed634fe037e79cb9315bc5b30df9b0f731e290b3b87f9af9b75a9"} Dec 06 11:00:06 crc kubenswrapper[4678]: I1206 11:00:06.408472 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" podStartSLOduration=3.408453077 podStartE2EDuration="3.408453077s" podCreationTimestamp="2025-12-06 11:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:00:06.404972105 +0000 UTC m=+1411.248403544" watchObservedRunningTime="2025-12-06 11:00:06.408453077 +0000 UTC m=+1411.251884516" Dec 06 11:00:07 crc kubenswrapper[4678]: I1206 11:00:07.397145 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8","Type":"ContainerStarted","Data":"43788aa981e063082a8a56cbd7cbc0d7add762ec2fbe36a7a4c85171493650bf"} Dec 06 11:00:07 crc kubenswrapper[4678]: I1206 11:00:07.398850 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3","Type":"ContainerStarted","Data":"6cb481599c4f8c22c4d72d2d23fe2488e7bc484b53203f39a380553cda96d28a"} Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.023807 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.113162 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-ld2rs"] Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.113472 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" podUID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerName="dnsmasq-dns" containerID="cri-o://ea562793da98cd8d0e1a3624820e2b17ae7f6e8fae92eeb224d5f48b247ecf3e" gracePeriod=10 Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.301098 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67cb876dc9-5749h"] Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.303395 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.321678 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67cb876dc9-5749h"] Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.433842 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-dns-swift-storage-0\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.433879 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-openstack-edpm-ipam\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.433925 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-config\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.433978 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-ovsdbserver-sb\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.434005 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j5md\" (UniqueName: \"kubernetes.io/projected/1ab593af-6400-4c78-b798-fd34038f28dd-kube-api-access-2j5md\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.434043 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-dns-svc\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.434220 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-ovsdbserver-nb\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.478744 4678 generic.go:334] "Generic (PLEG): container finished" podID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerID="ea562793da98cd8d0e1a3624820e2b17ae7f6e8fae92eeb224d5f48b247ecf3e" exitCode=0 Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.478790 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" event={"ID":"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1","Type":"ContainerDied","Data":"ea562793da98cd8d0e1a3624820e2b17ae7f6e8fae92eeb224d5f48b247ecf3e"} Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.538584 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-dns-swift-storage-0\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.538886 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-openstack-edpm-ipam\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.538933 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-config\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.538986 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-ovsdbserver-sb\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.539017 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j5md\" (UniqueName: \"kubernetes.io/projected/1ab593af-6400-4c78-b798-fd34038f28dd-kube-api-access-2j5md\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.539065 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-dns-svc\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.539099 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-ovsdbserver-nb\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.539423 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-dns-swift-storage-0\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.540739 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-ovsdbserver-nb\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.540852 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-openstack-edpm-ipam\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.541613 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-config\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.542635 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-ovsdbserver-sb\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.545173 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ab593af-6400-4c78-b798-fd34038f28dd-dns-svc\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.560148 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j5md\" (UniqueName: \"kubernetes.io/projected/1ab593af-6400-4c78-b798-fd34038f28dd-kube-api-access-2j5md\") pod \"dnsmasq-dns-67cb876dc9-5749h\" (UID: \"1ab593af-6400-4c78-b798-fd34038f28dd\") " pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.632747 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.641870 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.753021 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxvbs\" (UniqueName: \"kubernetes.io/projected/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-kube-api-access-gxvbs\") pod \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.753114 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-svc\") pod \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.753194 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-sb\") pod \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.753216 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-swift-storage-0\") pod \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.753269 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-config\") pod \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.753316 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-nb\") pod \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\" (UID: \"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1\") " Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.762437 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-kube-api-access-gxvbs" (OuterVolumeSpecName: "kube-api-access-gxvbs") pod "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" (UID: "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1"). InnerVolumeSpecName "kube-api-access-gxvbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.820716 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" (UID: "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:14 crc kubenswrapper[4678]: I1206 11:00:14.825717 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" (UID: "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.835230 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" (UID: "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.837826 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" (UID: "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.847443 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-config" (OuterVolumeSpecName: "config") pod "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" (UID: "dcd3fd07-7e77-48da-b5e0-e842e3acc7b1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.855783 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.855806 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxvbs\" (UniqueName: \"kubernetes.io/projected/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-kube-api-access-gxvbs\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.855820 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.855829 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.855838 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:14.855846 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1-config\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.128602 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67cb876dc9-5749h"] Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.495400 4678 generic.go:334] "Generic (PLEG): container finished" podID="1ab593af-6400-4c78-b798-fd34038f28dd" containerID="b28fddd7d76adba0ce6b3f9ecb0c3bc11d5b929257a03590c74c82cede074ceb" exitCode=0 Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.495470 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" event={"ID":"1ab593af-6400-4c78-b798-fd34038f28dd","Type":"ContainerDied","Data":"b28fddd7d76adba0ce6b3f9ecb0c3bc11d5b929257a03590c74c82cede074ceb"} Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.495849 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" event={"ID":"1ab593af-6400-4c78-b798-fd34038f28dd","Type":"ContainerStarted","Data":"6641ace75d62d1bc6ef72d595aa9c1a16747f6fba53c21e4e6931105bc1ab856"} Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.497794 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" event={"ID":"dcd3fd07-7e77-48da-b5e0-e842e3acc7b1","Type":"ContainerDied","Data":"b3b4767b0a2aefef7cf7371efaaef6cf69654509cf13c367f51280fcc889f8c4"} Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.497852 4678 scope.go:117] "RemoveContainer" containerID="ea562793da98cd8d0e1a3624820e2b17ae7f6e8fae92eeb224d5f48b247ecf3e" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.498050 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-ld2rs" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.706410 4678 scope.go:117] "RemoveContainer" containerID="5e84ff4077e11d33c2dc981d1419d793f408cb1c405425671042459174bf04f4" Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.738470 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-ld2rs"] Dec 06 11:00:15 crc kubenswrapper[4678]: I1206 11:00:15.746983 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-ld2rs"] Dec 06 11:00:16 crc kubenswrapper[4678]: I1206 11:00:16.507730 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" event={"ID":"1ab593af-6400-4c78-b798-fd34038f28dd","Type":"ContainerStarted","Data":"d5e2f9598639feedfd5468182a9142f6c202d61f74a6e33cc9f663b2ea9428f8"} Dec 06 11:00:16 crc kubenswrapper[4678]: I1206 11:00:16.532205 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" podStartSLOduration=2.532187063 podStartE2EDuration="2.532187063s" podCreationTimestamp="2025-12-06 11:00:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:00:16.525753133 +0000 UTC m=+1421.369184572" watchObservedRunningTime="2025-12-06 11:00:16.532187063 +0000 UTC m=+1421.375618512" Dec 06 11:00:17 crc kubenswrapper[4678]: I1206 11:00:17.487559 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" path="/var/lib/kubelet/pods/dcd3fd07-7e77-48da-b5e0-e842e3acc7b1/volumes" Dec 06 11:00:17 crc kubenswrapper[4678]: I1206 11:00:17.519668 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.612956 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-26z9r"] Dec 06 11:00:21 crc kubenswrapper[4678]: E1206 11:00:21.615511 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerName="dnsmasq-dns" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.615622 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerName="dnsmasq-dns" Dec 06 11:00:21 crc kubenswrapper[4678]: E1206 11:00:21.615729 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerName="init" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.615821 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerName="init" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.616176 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcd3fd07-7e77-48da-b5e0-e842e3acc7b1" containerName="dnsmasq-dns" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.618126 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.624290 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-26z9r"] Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.717242 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-utilities\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.717395 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-catalog-content\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.717690 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npkln\" (UniqueName: \"kubernetes.io/projected/a29dd839-09ba-44f8-a0b6-c874b6e427dc-kube-api-access-npkln\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.819395 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-utilities\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.819643 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-catalog-content\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.819814 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npkln\" (UniqueName: \"kubernetes.io/projected/a29dd839-09ba-44f8-a0b6-c874b6e427dc-kube-api-access-npkln\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.820011 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-catalog-content\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.820205 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-utilities\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.855449 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npkln\" (UniqueName: \"kubernetes.io/projected/a29dd839-09ba-44f8-a0b6-c874b6e427dc-kube-api-access-npkln\") pod \"redhat-operators-26z9r\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:21 crc kubenswrapper[4678]: I1206 11:00:21.938353 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:22 crc kubenswrapper[4678]: I1206 11:00:22.479079 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-26z9r"] Dec 06 11:00:22 crc kubenswrapper[4678]: I1206 11:00:22.572333 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26z9r" event={"ID":"a29dd839-09ba-44f8-a0b6-c874b6e427dc","Type":"ContainerStarted","Data":"d582f5786da36fd220d62eb0c4d9ed47fc87cba257950e227807907cebaca756"} Dec 06 11:00:23 crc kubenswrapper[4678]: I1206 11:00:23.584460 4678 generic.go:334] "Generic (PLEG): container finished" podID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerID="64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7" exitCode=0 Dec 06 11:00:23 crc kubenswrapper[4678]: I1206 11:00:23.584524 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26z9r" event={"ID":"a29dd839-09ba-44f8-a0b6-c874b6e427dc","Type":"ContainerDied","Data":"64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7"} Dec 06 11:00:24 crc kubenswrapper[4678]: I1206 11:00:24.643370 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67cb876dc9-5749h" Dec 06 11:00:24 crc kubenswrapper[4678]: I1206 11:00:24.718032 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-4zcbf"] Dec 06 11:00:24 crc kubenswrapper[4678]: I1206 11:00:24.718294 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" podUID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerName="dnsmasq-dns" containerID="cri-o://3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199" gracePeriod=10 Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.253613 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.285707 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-nb\") pod \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.286140 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-svc\") pod \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.286171 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p42qt\" (UniqueName: \"kubernetes.io/projected/4dac7d51-741a-418d-8eb4-e4583fe9cc63-kube-api-access-p42qt\") pod \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.309468 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dac7d51-741a-418d-8eb4-e4583fe9cc63-kube-api-access-p42qt" (OuterVolumeSpecName: "kube-api-access-p42qt") pod "4dac7d51-741a-418d-8eb4-e4583fe9cc63" (UID: "4dac7d51-741a-418d-8eb4-e4583fe9cc63"). InnerVolumeSpecName "kube-api-access-p42qt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.388530 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-sb\") pod \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.388785 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-config\") pod \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.388828 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-openstack-edpm-ipam\") pod \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.388947 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-swift-storage-0\") pod \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\" (UID: \"4dac7d51-741a-418d-8eb4-e4583fe9cc63\") " Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.391082 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p42qt\" (UniqueName: \"kubernetes.io/projected/4dac7d51-741a-418d-8eb4-e4583fe9cc63-kube-api-access-p42qt\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.436460 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4dac7d51-741a-418d-8eb4-e4583fe9cc63" (UID: "4dac7d51-741a-418d-8eb4-e4583fe9cc63"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.438245 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4dac7d51-741a-418d-8eb4-e4583fe9cc63" (UID: "4dac7d51-741a-418d-8eb4-e4583fe9cc63"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.445396 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4dac7d51-741a-418d-8eb4-e4583fe9cc63" (UID: "4dac7d51-741a-418d-8eb4-e4583fe9cc63"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.474255 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4dac7d51-741a-418d-8eb4-e4583fe9cc63" (UID: "4dac7d51-741a-418d-8eb4-e4583fe9cc63"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.483975 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "4dac7d51-741a-418d-8eb4-e4583fe9cc63" (UID: "4dac7d51-741a-418d-8eb4-e4583fe9cc63"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.493102 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.493139 4678 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.493152 4678 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.493163 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.493178 4678 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.502873 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-config" (OuterVolumeSpecName: "config") pod "4dac7d51-741a-418d-8eb4-e4583fe9cc63" (UID: "4dac7d51-741a-418d-8eb4-e4583fe9cc63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.594911 4678 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dac7d51-741a-418d-8eb4-e4583fe9cc63-config\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.607921 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26z9r" event={"ID":"a29dd839-09ba-44f8-a0b6-c874b6e427dc","Type":"ContainerStarted","Data":"ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27"} Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.613038 4678 generic.go:334] "Generic (PLEG): container finished" podID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerID="3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199" exitCode=0 Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.613089 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" event={"ID":"4dac7d51-741a-418d-8eb4-e4583fe9cc63","Type":"ContainerDied","Data":"3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199"} Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.613114 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.613139 4678 scope.go:117] "RemoveContainer" containerID="3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.613126 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-4zcbf" event={"ID":"4dac7d51-741a-418d-8eb4-e4583fe9cc63","Type":"ContainerDied","Data":"0770d47c7b9dbe492e9dc9b4024262cfc4f084d7b99a2eab00ac9cf185668e66"} Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.654073 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-4zcbf"] Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.657517 4678 scope.go:117] "RemoveContainer" containerID="534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.662555 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-4zcbf"] Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.677244 4678 scope.go:117] "RemoveContainer" containerID="3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199" Dec 06 11:00:25 crc kubenswrapper[4678]: E1206 11:00:25.677908 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199\": container with ID starting with 3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199 not found: ID does not exist" containerID="3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.677973 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199"} err="failed to get container status \"3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199\": rpc error: code = NotFound desc = could not find container \"3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199\": container with ID starting with 3a2ae9f27f96804a68214d19b5febe89fc948f83bc3970bdd7e5bab64f36b199 not found: ID does not exist" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.678015 4678 scope.go:117] "RemoveContainer" containerID="534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d" Dec 06 11:00:25 crc kubenswrapper[4678]: E1206 11:00:25.678438 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d\": container with ID starting with 534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d not found: ID does not exist" containerID="534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d" Dec 06 11:00:25 crc kubenswrapper[4678]: I1206 11:00:25.678534 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d"} err="failed to get container status \"534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d\": rpc error: code = NotFound desc = could not find container \"534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d\": container with ID starting with 534a68c238d15b33121ce49a184dda010a32120e873fe1756c23c0063129e74d not found: ID does not exist" Dec 06 11:00:27 crc kubenswrapper[4678]: I1206 11:00:27.504857 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" path="/var/lib/kubelet/pods/4dac7d51-741a-418d-8eb4-e4583fe9cc63/volumes" Dec 06 11:00:29 crc kubenswrapper[4678]: I1206 11:00:29.656279 4678 generic.go:334] "Generic (PLEG): container finished" podID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerID="ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27" exitCode=0 Dec 06 11:00:29 crc kubenswrapper[4678]: I1206 11:00:29.656554 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26z9r" event={"ID":"a29dd839-09ba-44f8-a0b6-c874b6e427dc","Type":"ContainerDied","Data":"ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27"} Dec 06 11:00:30 crc kubenswrapper[4678]: I1206 11:00:30.667232 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26z9r" event={"ID":"a29dd839-09ba-44f8-a0b6-c874b6e427dc","Type":"ContainerStarted","Data":"b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693"} Dec 06 11:00:30 crc kubenswrapper[4678]: I1206 11:00:30.694474 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-26z9r" podStartSLOduration=3.197602205 podStartE2EDuration="9.694456109s" podCreationTimestamp="2025-12-06 11:00:21 +0000 UTC" firstStartedPulling="2025-12-06 11:00:23.587260269 +0000 UTC m=+1428.430691748" lastFinishedPulling="2025-12-06 11:00:30.084114213 +0000 UTC m=+1434.927545652" observedRunningTime="2025-12-06 11:00:30.683607576 +0000 UTC m=+1435.527039025" watchObservedRunningTime="2025-12-06 11:00:30.694456109 +0000 UTC m=+1435.537887538" Dec 06 11:00:31 crc kubenswrapper[4678]: I1206 11:00:31.938778 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:31 crc kubenswrapper[4678]: I1206 11:00:31.938839 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:32 crc kubenswrapper[4678]: I1206 11:00:32.984915 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-26z9r" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="registry-server" probeResult="failure" output=< Dec 06 11:00:32 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:00:32 crc kubenswrapper[4678]: > Dec 06 11:00:39 crc kubenswrapper[4678]: I1206 11:00:39.752080 4678 generic.go:334] "Generic (PLEG): container finished" podID="cb4272e8-bb6c-43d4-8539-baeb4f3c14a8" containerID="43788aa981e063082a8a56cbd7cbc0d7add762ec2fbe36a7a4c85171493650bf" exitCode=0 Dec 06 11:00:39 crc kubenswrapper[4678]: I1206 11:00:39.752287 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8","Type":"ContainerDied","Data":"43788aa981e063082a8a56cbd7cbc0d7add762ec2fbe36a7a4c85171493650bf"} Dec 06 11:00:39 crc kubenswrapper[4678]: I1206 11:00:39.759382 4678 generic.go:334] "Generic (PLEG): container finished" podID="f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3" containerID="6cb481599c4f8c22c4d72d2d23fe2488e7bc484b53203f39a380553cda96d28a" exitCode=0 Dec 06 11:00:39 crc kubenswrapper[4678]: I1206 11:00:39.759428 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3","Type":"ContainerDied","Data":"6cb481599c4f8c22c4d72d2d23fe2488e7bc484b53203f39a380553cda96d28a"} Dec 06 11:00:40 crc kubenswrapper[4678]: I1206 11:00:40.771582 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3","Type":"ContainerStarted","Data":"53b9bbe612642070dc63b79ee2f99a49a10604a7241b6d5c9a3e893e448f419e"} Dec 06 11:00:40 crc kubenswrapper[4678]: I1206 11:00:40.772298 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:00:40 crc kubenswrapper[4678]: I1206 11:00:40.774268 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"cb4272e8-bb6c-43d4-8539-baeb4f3c14a8","Type":"ContainerStarted","Data":"67c697aeb372dc6bc57f5d83f89feffe36a5e61ff8cf9445c0e9b94a4e99e617"} Dec 06 11:00:40 crc kubenswrapper[4678]: I1206 11:00:40.774762 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 06 11:00:40 crc kubenswrapper[4678]: I1206 11:00:40.800980 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.800958615 podStartE2EDuration="36.800958615s" podCreationTimestamp="2025-12-06 11:00:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:00:40.790569397 +0000 UTC m=+1445.634000846" watchObservedRunningTime="2025-12-06 11:00:40.800958615 +0000 UTC m=+1445.644390074" Dec 06 11:00:40 crc kubenswrapper[4678]: I1206 11:00:40.844085 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.844065286 podStartE2EDuration="37.844065286s" podCreationTimestamp="2025-12-06 11:00:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:00:40.835449436 +0000 UTC m=+1445.678880885" watchObservedRunningTime="2025-12-06 11:00:40.844065286 +0000 UTC m=+1445.687496745" Dec 06 11:00:41 crc kubenswrapper[4678]: I1206 11:00:41.999122 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:42 crc kubenswrapper[4678]: I1206 11:00:42.070416 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:42 crc kubenswrapper[4678]: I1206 11:00:42.250406 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-26z9r"] Dec 06 11:00:43 crc kubenswrapper[4678]: I1206 11:00:43.805303 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-26z9r" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="registry-server" containerID="cri-o://b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693" gracePeriod=2 Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.251567 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.272152 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-utilities\") pod \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.272316 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npkln\" (UniqueName: \"kubernetes.io/projected/a29dd839-09ba-44f8-a0b6-c874b6e427dc-kube-api-access-npkln\") pod \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.272427 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-catalog-content\") pod \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\" (UID: \"a29dd839-09ba-44f8-a0b6-c874b6e427dc\") " Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.272849 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-utilities" (OuterVolumeSpecName: "utilities") pod "a29dd839-09ba-44f8-a0b6-c874b6e427dc" (UID: "a29dd839-09ba-44f8-a0b6-c874b6e427dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.293533 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a29dd839-09ba-44f8-a0b6-c874b6e427dc-kube-api-access-npkln" (OuterVolumeSpecName: "kube-api-access-npkln") pod "a29dd839-09ba-44f8-a0b6-c874b6e427dc" (UID: "a29dd839-09ba-44f8-a0b6-c874b6e427dc"). InnerVolumeSpecName "kube-api-access-npkln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.309891 4678 scope.go:117] "RemoveContainer" containerID="0b3242f24bd29cf8c95a52f2852b84550f148909fe5ac46f4627d3cd7805dbf6" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.343422 4678 scope.go:117] "RemoveContainer" containerID="2c94b4c0a6c2dbabaf6944ff038082789e635ac8c153e525d547a63b452a54bd" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.375026 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.375055 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npkln\" (UniqueName: \"kubernetes.io/projected/a29dd839-09ba-44f8-a0b6-c874b6e427dc-kube-api-access-npkln\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.399857 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a29dd839-09ba-44f8-a0b6-c874b6e427dc" (UID: "a29dd839-09ba-44f8-a0b6-c874b6e427dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.476849 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a29dd839-09ba-44f8-a0b6-c874b6e427dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.819170 4678 generic.go:334] "Generic (PLEG): container finished" podID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerID="b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693" exitCode=0 Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.819219 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26z9r" event={"ID":"a29dd839-09ba-44f8-a0b6-c874b6e427dc","Type":"ContainerDied","Data":"b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693"} Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.819251 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-26z9r" event={"ID":"a29dd839-09ba-44f8-a0b6-c874b6e427dc","Type":"ContainerDied","Data":"d582f5786da36fd220d62eb0c4d9ed47fc87cba257950e227807907cebaca756"} Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.819247 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-26z9r" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.819269 4678 scope.go:117] "RemoveContainer" containerID="b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.841452 4678 scope.go:117] "RemoveContainer" containerID="ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.865969 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-26z9r"] Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.875623 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-26z9r"] Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.880875 4678 scope.go:117] "RemoveContainer" containerID="64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.900051 4678 scope.go:117] "RemoveContainer" containerID="b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693" Dec 06 11:00:44 crc kubenswrapper[4678]: E1206 11:00:44.900454 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693\": container with ID starting with b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693 not found: ID does not exist" containerID="b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.900519 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693"} err="failed to get container status \"b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693\": rpc error: code = NotFound desc = could not find container \"b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693\": container with ID starting with b5ca43406dd00d98c395e51e861717a657e1caf553c8e25c85613ef22a981693 not found: ID does not exist" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.900545 4678 scope.go:117] "RemoveContainer" containerID="ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27" Dec 06 11:00:44 crc kubenswrapper[4678]: E1206 11:00:44.900818 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27\": container with ID starting with ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27 not found: ID does not exist" containerID="ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.900857 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27"} err="failed to get container status \"ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27\": rpc error: code = NotFound desc = could not find container \"ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27\": container with ID starting with ee7f951f633986ec1a8d4da624196983a883b8078e696f30e5db060387672f27 not found: ID does not exist" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.900878 4678 scope.go:117] "RemoveContainer" containerID="64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7" Dec 06 11:00:44 crc kubenswrapper[4678]: E1206 11:00:44.901148 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7\": container with ID starting with 64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7 not found: ID does not exist" containerID="64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7" Dec 06 11:00:44 crc kubenswrapper[4678]: I1206 11:00:44.901176 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7"} err="failed to get container status \"64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7\": rpc error: code = NotFound desc = could not find container \"64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7\": container with ID starting with 64cc6f822fe910f7ad10c2badbba4ef3de2f119c268add364c259a20e4ad6ff7 not found: ID does not exist" Dec 06 11:00:44 crc kubenswrapper[4678]: E1206 11:00:44.959181 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda29dd839_09ba_44f8_a0b6_c874b6e427dc.slice/crio-d582f5786da36fd220d62eb0c4d9ed47fc87cba257950e227807907cebaca756\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda29dd839_09ba_44f8_a0b6_c874b6e427dc.slice\": RecentStats: unable to find data in memory cache]" Dec 06 11:00:45 crc kubenswrapper[4678]: I1206 11:00:45.520665 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" path="/var/lib/kubelet/pods/a29dd839-09ba-44f8-a0b6-c874b6e427dc/volumes" Dec 06 11:00:45 crc kubenswrapper[4678]: I1206 11:00:45.687757 4678 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poddcd3fd07-7e77-48da-b5e0-e842e3acc7b1"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poddcd3fd07-7e77-48da-b5e0-e842e3acc7b1] : Timed out while waiting for systemd to remove kubepods-besteffort-poddcd3fd07_7e77_48da_b5e0_e842e3acc7b1.slice" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.938340 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb"] Dec 06 11:00:47 crc kubenswrapper[4678]: E1206 11:00:47.938859 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="registry-server" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.938877 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="registry-server" Dec 06 11:00:47 crc kubenswrapper[4678]: E1206 11:00:47.938910 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="extract-utilities" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.938919 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="extract-utilities" Dec 06 11:00:47 crc kubenswrapper[4678]: E1206 11:00:47.938933 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerName="init" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.938942 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerName="init" Dec 06 11:00:47 crc kubenswrapper[4678]: E1206 11:00:47.938972 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="extract-content" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.938979 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="extract-content" Dec 06 11:00:47 crc kubenswrapper[4678]: E1206 11:00:47.938996 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerName="dnsmasq-dns" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.939003 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerName="dnsmasq-dns" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.939260 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a29dd839-09ba-44f8-a0b6-c874b6e427dc" containerName="registry-server" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.939279 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dac7d51-741a-418d-8eb4-e4583fe9cc63" containerName="dnsmasq-dns" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.940041 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.942402 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.942691 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.943388 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.943617 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:00:47 crc kubenswrapper[4678]: I1206 11:00:47.956639 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb"] Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.041069 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.041364 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.041392 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.041414 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54l49\" (UniqueName: \"kubernetes.io/projected/29b633cb-d39d-43b9-9556-461cd563c92f-kube-api-access-54l49\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.143317 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.143382 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.143425 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54l49\" (UniqueName: \"kubernetes.io/projected/29b633cb-d39d-43b9-9556-461cd563c92f-kube-api-access-54l49\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.143597 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.149982 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.150073 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.150679 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.178807 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54l49\" (UniqueName: \"kubernetes.io/projected/29b633cb-d39d-43b9-9556-461cd563c92f-kube-api-access-54l49\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.269143 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:00:48 crc kubenswrapper[4678]: W1206 11:00:48.936112 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29b633cb_d39d_43b9_9556_461cd563c92f.slice/crio-f662feda05714c89e6fe654c42a46baccae98fbef56d9216282f3e0f89da8974 WatchSource:0}: Error finding container f662feda05714c89e6fe654c42a46baccae98fbef56d9216282f3e0f89da8974: Status 404 returned error can't find the container with id f662feda05714c89e6fe654c42a46baccae98fbef56d9216282f3e0f89da8974 Dec 06 11:00:48 crc kubenswrapper[4678]: I1206 11:00:48.936881 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb"] Dec 06 11:00:49 crc kubenswrapper[4678]: I1206 11:00:49.874239 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" event={"ID":"29b633cb-d39d-43b9-9556-461cd563c92f","Type":"ContainerStarted","Data":"f662feda05714c89e6fe654c42a46baccae98fbef56d9216282f3e0f89da8974"} Dec 06 11:00:54 crc kubenswrapper[4678]: I1206 11:00:54.319290 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 06 11:00:55 crc kubenswrapper[4678]: I1206 11:00:55.335666 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.153335 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29416981-627fr"] Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.158507 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.168209 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416981-627fr"] Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.308414 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-config-data\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.308743 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svj28\" (UniqueName: \"kubernetes.io/projected/29f472e5-54ad-470c-a43b-a6fbee7fa24d-kube-api-access-svj28\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.308821 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-combined-ca-bundle\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.309096 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-fernet-keys\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.410747 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-combined-ca-bundle\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.410864 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-fernet-keys\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.411262 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-config-data\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.411305 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svj28\" (UniqueName: \"kubernetes.io/projected/29f472e5-54ad-470c-a43b-a6fbee7fa24d-kube-api-access-svj28\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.418081 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-config-data\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.424881 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-fernet-keys\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.426766 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svj28\" (UniqueName: \"kubernetes.io/projected/29f472e5-54ad-470c-a43b-a6fbee7fa24d-kube-api-access-svj28\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.427256 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-combined-ca-bundle\") pod \"keystone-cron-29416981-627fr\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.488436 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:00 crc kubenswrapper[4678]: I1206 11:01:00.971359 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29416981-627fr"] Dec 06 11:01:01 crc kubenswrapper[4678]: I1206 11:01:01.018587 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" event={"ID":"29b633cb-d39d-43b9-9556-461cd563c92f","Type":"ContainerStarted","Data":"2eb941859a9121145f1e03c90a894b41660c35c2e026bdd02d2164fb9747dfb9"} Dec 06 11:01:01 crc kubenswrapper[4678]: I1206 11:01:01.020778 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416981-627fr" event={"ID":"29f472e5-54ad-470c-a43b-a6fbee7fa24d","Type":"ContainerStarted","Data":"60d13fb1606df0ad4946171bf4645a4746d10b371ae0a6b5443d7305c6c51b67"} Dec 06 11:01:01 crc kubenswrapper[4678]: I1206 11:01:01.047943 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" podStartSLOduration=2.573193969 podStartE2EDuration="14.047927631s" podCreationTimestamp="2025-12-06 11:00:47 +0000 UTC" firstStartedPulling="2025-12-06 11:00:48.938516462 +0000 UTC m=+1453.781947911" lastFinishedPulling="2025-12-06 11:01:00.413250134 +0000 UTC m=+1465.256681573" observedRunningTime="2025-12-06 11:01:01.046898548 +0000 UTC m=+1465.890329987" watchObservedRunningTime="2025-12-06 11:01:01.047927631 +0000 UTC m=+1465.891359070" Dec 06 11:01:02 crc kubenswrapper[4678]: I1206 11:01:02.029395 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416981-627fr" event={"ID":"29f472e5-54ad-470c-a43b-a6fbee7fa24d","Type":"ContainerStarted","Data":"167af0c5e5422ef1249e74d5b3a21d06306d8c8ba574ab05e5049f06327c0a22"} Dec 06 11:01:02 crc kubenswrapper[4678]: I1206 11:01:02.053169 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29416981-627fr" podStartSLOduration=2.05314804 podStartE2EDuration="2.05314804s" podCreationTimestamp="2025-12-06 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:01:02.043230608 +0000 UTC m=+1466.886662057" watchObservedRunningTime="2025-12-06 11:01:02.05314804 +0000 UTC m=+1466.896579479" Dec 06 11:01:05 crc kubenswrapper[4678]: I1206 11:01:05.058974 4678 generic.go:334] "Generic (PLEG): container finished" podID="29f472e5-54ad-470c-a43b-a6fbee7fa24d" containerID="167af0c5e5422ef1249e74d5b3a21d06306d8c8ba574ab05e5049f06327c0a22" exitCode=0 Dec 06 11:01:05 crc kubenswrapper[4678]: I1206 11:01:05.059048 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416981-627fr" event={"ID":"29f472e5-54ad-470c-a43b-a6fbee7fa24d","Type":"ContainerDied","Data":"167af0c5e5422ef1249e74d5b3a21d06306d8c8ba574ab05e5049f06327c0a22"} Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.553966 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.638356 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-config-data\") pod \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.638670 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svj28\" (UniqueName: \"kubernetes.io/projected/29f472e5-54ad-470c-a43b-a6fbee7fa24d-kube-api-access-svj28\") pod \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.638806 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-fernet-keys\") pod \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.638945 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-combined-ca-bundle\") pod \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\" (UID: \"29f472e5-54ad-470c-a43b-a6fbee7fa24d\") " Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.650631 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29f472e5-54ad-470c-a43b-a6fbee7fa24d-kube-api-access-svj28" (OuterVolumeSpecName: "kube-api-access-svj28") pod "29f472e5-54ad-470c-a43b-a6fbee7fa24d" (UID: "29f472e5-54ad-470c-a43b-a6fbee7fa24d"). InnerVolumeSpecName "kube-api-access-svj28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.680656 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "29f472e5-54ad-470c-a43b-a6fbee7fa24d" (UID: "29f472e5-54ad-470c-a43b-a6fbee7fa24d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.726202 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29f472e5-54ad-470c-a43b-a6fbee7fa24d" (UID: "29f472e5-54ad-470c-a43b-a6fbee7fa24d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.741316 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svj28\" (UniqueName: \"kubernetes.io/projected/29f472e5-54ad-470c-a43b-a6fbee7fa24d-kube-api-access-svj28\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.741353 4678 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.741367 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.768507 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-config-data" (OuterVolumeSpecName: "config-data") pod "29f472e5-54ad-470c-a43b-a6fbee7fa24d" (UID: "29f472e5-54ad-470c-a43b-a6fbee7fa24d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:06 crc kubenswrapper[4678]: I1206 11:01:06.842702 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29f472e5-54ad-470c-a43b-a6fbee7fa24d-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:07 crc kubenswrapper[4678]: I1206 11:01:07.220216 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29416981-627fr" event={"ID":"29f472e5-54ad-470c-a43b-a6fbee7fa24d","Type":"ContainerDied","Data":"60d13fb1606df0ad4946171bf4645a4746d10b371ae0a6b5443d7305c6c51b67"} Dec 06 11:01:07 crc kubenswrapper[4678]: I1206 11:01:07.220256 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60d13fb1606df0ad4946171bf4645a4746d10b371ae0a6b5443d7305c6c51b67" Dec 06 11:01:07 crc kubenswrapper[4678]: I1206 11:01:07.220315 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29416981-627fr" Dec 06 11:01:12 crc kubenswrapper[4678]: I1206 11:01:12.269175 4678 generic.go:334] "Generic (PLEG): container finished" podID="29b633cb-d39d-43b9-9556-461cd563c92f" containerID="2eb941859a9121145f1e03c90a894b41660c35c2e026bdd02d2164fb9747dfb9" exitCode=0 Dec 06 11:01:12 crc kubenswrapper[4678]: I1206 11:01:12.269228 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" event={"ID":"29b633cb-d39d-43b9-9556-461cd563c92f","Type":"ContainerDied","Data":"2eb941859a9121145f1e03c90a894b41660c35c2e026bdd02d2164fb9747dfb9"} Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.687155 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.784095 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-ssh-key\") pod \"29b633cb-d39d-43b9-9556-461cd563c92f\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.784342 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54l49\" (UniqueName: \"kubernetes.io/projected/29b633cb-d39d-43b9-9556-461cd563c92f-kube-api-access-54l49\") pod \"29b633cb-d39d-43b9-9556-461cd563c92f\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.784386 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-repo-setup-combined-ca-bundle\") pod \"29b633cb-d39d-43b9-9556-461cd563c92f\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.784409 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-inventory\") pod \"29b633cb-d39d-43b9-9556-461cd563c92f\" (UID: \"29b633cb-d39d-43b9-9556-461cd563c92f\") " Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.815724 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29b633cb-d39d-43b9-9556-461cd563c92f-kube-api-access-54l49" (OuterVolumeSpecName: "kube-api-access-54l49") pod "29b633cb-d39d-43b9-9556-461cd563c92f" (UID: "29b633cb-d39d-43b9-9556-461cd563c92f"). InnerVolumeSpecName "kube-api-access-54l49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.816288 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "29b633cb-d39d-43b9-9556-461cd563c92f" (UID: "29b633cb-d39d-43b9-9556-461cd563c92f"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.850651 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-inventory" (OuterVolumeSpecName: "inventory") pod "29b633cb-d39d-43b9-9556-461cd563c92f" (UID: "29b633cb-d39d-43b9-9556-461cd563c92f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.886685 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54l49\" (UniqueName: \"kubernetes.io/projected/29b633cb-d39d-43b9-9556-461cd563c92f-kube-api-access-54l49\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.887355 4678 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.887431 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.915632 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "29b633cb-d39d-43b9-9556-461cd563c92f" (UID: "29b633cb-d39d-43b9-9556-461cd563c92f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:13 crc kubenswrapper[4678]: I1206 11:01:13.989409 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/29b633cb-d39d-43b9-9556-461cd563c92f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.287866 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" event={"ID":"29b633cb-d39d-43b9-9556-461cd563c92f","Type":"ContainerDied","Data":"f662feda05714c89e6fe654c42a46baccae98fbef56d9216282f3e0f89da8974"} Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.287908 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f662feda05714c89e6fe654c42a46baccae98fbef56d9216282f3e0f89da8974" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.288242 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.434138 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh"] Dec 06 11:01:14 crc kubenswrapper[4678]: E1206 11:01:14.434711 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29f472e5-54ad-470c-a43b-a6fbee7fa24d" containerName="keystone-cron" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.434735 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="29f472e5-54ad-470c-a43b-a6fbee7fa24d" containerName="keystone-cron" Dec 06 11:01:14 crc kubenswrapper[4678]: E1206 11:01:14.434776 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b633cb-d39d-43b9-9556-461cd563c92f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.434791 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b633cb-d39d-43b9-9556-461cd563c92f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.435053 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="29f472e5-54ad-470c-a43b-a6fbee7fa24d" containerName="keystone-cron" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.435086 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="29b633cb-d39d-43b9-9556-461cd563c92f" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.435857 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.439231 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.439401 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.439596 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.439705 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.456676 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh"] Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.601976 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.602175 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.602237 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4mx2\" (UniqueName: \"kubernetes.io/projected/d66737d1-caae-4204-9786-2b19a88f5f82-kube-api-access-p4mx2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.704192 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.704313 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.704484 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4mx2\" (UniqueName: \"kubernetes.io/projected/d66737d1-caae-4204-9786-2b19a88f5f82-kube-api-access-p4mx2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.708261 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.712366 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.738126 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4mx2\" (UniqueName: \"kubernetes.io/projected/d66737d1-caae-4204-9786-2b19a88f5f82-kube-api-access-p4mx2\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-jb8nh\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:14 crc kubenswrapper[4678]: I1206 11:01:14.760460 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:15 crc kubenswrapper[4678]: I1206 11:01:15.340741 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh"] Dec 06 11:01:16 crc kubenswrapper[4678]: I1206 11:01:16.307418 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" event={"ID":"d66737d1-caae-4204-9786-2b19a88f5f82","Type":"ContainerStarted","Data":"fdd02df0849aabb5cf9cbe1cdeb47ad9aba20527d821368f3cec23eb01f05233"} Dec 06 11:01:16 crc kubenswrapper[4678]: I1206 11:01:16.307932 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" event={"ID":"d66737d1-caae-4204-9786-2b19a88f5f82","Type":"ContainerStarted","Data":"456a27824a56c2653226b26dc69d42a5e8718820bfd18aff7f73c9494eb50ca8"} Dec 06 11:01:16 crc kubenswrapper[4678]: I1206 11:01:16.329339 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" podStartSLOduration=1.7358124990000001 podStartE2EDuration="2.329319968s" podCreationTimestamp="2025-12-06 11:01:14 +0000 UTC" firstStartedPulling="2025-12-06 11:01:15.360595705 +0000 UTC m=+1480.204027144" lastFinishedPulling="2025-12-06 11:01:15.954103174 +0000 UTC m=+1480.797534613" observedRunningTime="2025-12-06 11:01:16.328702718 +0000 UTC m=+1481.172134177" watchObservedRunningTime="2025-12-06 11:01:16.329319968 +0000 UTC m=+1481.172751407" Dec 06 11:01:19 crc kubenswrapper[4678]: I1206 11:01:19.342138 4678 generic.go:334] "Generic (PLEG): container finished" podID="d66737d1-caae-4204-9786-2b19a88f5f82" containerID="fdd02df0849aabb5cf9cbe1cdeb47ad9aba20527d821368f3cec23eb01f05233" exitCode=0 Dec 06 11:01:19 crc kubenswrapper[4678]: I1206 11:01:19.342219 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" event={"ID":"d66737d1-caae-4204-9786-2b19a88f5f82","Type":"ContainerDied","Data":"fdd02df0849aabb5cf9cbe1cdeb47ad9aba20527d821368f3cec23eb01f05233"} Dec 06 11:01:20 crc kubenswrapper[4678]: I1206 11:01:20.821750 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:20 crc kubenswrapper[4678]: I1206 11:01:20.946652 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-ssh-key\") pod \"d66737d1-caae-4204-9786-2b19a88f5f82\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " Dec 06 11:01:20 crc kubenswrapper[4678]: I1206 11:01:20.946698 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4mx2\" (UniqueName: \"kubernetes.io/projected/d66737d1-caae-4204-9786-2b19a88f5f82-kube-api-access-p4mx2\") pod \"d66737d1-caae-4204-9786-2b19a88f5f82\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " Dec 06 11:01:20 crc kubenswrapper[4678]: I1206 11:01:20.946781 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-inventory\") pod \"d66737d1-caae-4204-9786-2b19a88f5f82\" (UID: \"d66737d1-caae-4204-9786-2b19a88f5f82\") " Dec 06 11:01:20 crc kubenswrapper[4678]: I1206 11:01:20.955230 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d66737d1-caae-4204-9786-2b19a88f5f82-kube-api-access-p4mx2" (OuterVolumeSpecName: "kube-api-access-p4mx2") pod "d66737d1-caae-4204-9786-2b19a88f5f82" (UID: "d66737d1-caae-4204-9786-2b19a88f5f82"). InnerVolumeSpecName "kube-api-access-p4mx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:01:20 crc kubenswrapper[4678]: I1206 11:01:20.978203 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d66737d1-caae-4204-9786-2b19a88f5f82" (UID: "d66737d1-caae-4204-9786-2b19a88f5f82"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:20 crc kubenswrapper[4678]: I1206 11:01:20.993441 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-inventory" (OuterVolumeSpecName: "inventory") pod "d66737d1-caae-4204-9786-2b19a88f5f82" (UID: "d66737d1-caae-4204-9786-2b19a88f5f82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.049143 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4mx2\" (UniqueName: \"kubernetes.io/projected/d66737d1-caae-4204-9786-2b19a88f5f82-kube-api-access-p4mx2\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.049544 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.049559 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d66737d1-caae-4204-9786-2b19a88f5f82-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.368606 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" event={"ID":"d66737d1-caae-4204-9786-2b19a88f5f82","Type":"ContainerDied","Data":"456a27824a56c2653226b26dc69d42a5e8718820bfd18aff7f73c9494eb50ca8"} Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.368682 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="456a27824a56c2653226b26dc69d42a5e8718820bfd18aff7f73c9494eb50ca8" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.368693 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-jb8nh" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.524982 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf"] Dec 06 11:01:21 crc kubenswrapper[4678]: E1206 11:01:21.525813 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66737d1-caae-4204-9786-2b19a88f5f82" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.525862 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66737d1-caae-4204-9786-2b19a88f5f82" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.526393 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d66737d1-caae-4204-9786-2b19a88f5f82" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.527571 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.534358 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.535882 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.537828 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.538033 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.565377 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf"] Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.675075 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.675158 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.675236 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.675391 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6c7b\" (UniqueName: \"kubernetes.io/projected/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-kube-api-access-q6c7b\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.777520 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6c7b\" (UniqueName: \"kubernetes.io/projected/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-kube-api-access-q6c7b\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.777628 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.777787 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.777866 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.782217 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.782351 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.785818 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.803128 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6c7b\" (UniqueName: \"kubernetes.io/projected/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-kube-api-access-q6c7b\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:21 crc kubenswrapper[4678]: I1206 11:01:21.848682 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:01:22 crc kubenswrapper[4678]: I1206 11:01:22.406097 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf"] Dec 06 11:01:23 crc kubenswrapper[4678]: I1206 11:01:23.389909 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" event={"ID":"99d806d5-73ca-4ecc-92cb-b3c300cb04b4","Type":"ContainerStarted","Data":"f6ab2f1feebb265a01d5f45c9e965958c1f768f78e9e7914c6b9d3f3fbc01208"} Dec 06 11:01:23 crc kubenswrapper[4678]: I1206 11:01:23.390227 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" event={"ID":"99d806d5-73ca-4ecc-92cb-b3c300cb04b4","Type":"ContainerStarted","Data":"47cc28df0829e7ded6349f3dfa8b4fd1d57c93fcf998e1d842a105178bfb8039"} Dec 06 11:01:23 crc kubenswrapper[4678]: I1206 11:01:23.416453 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" podStartSLOduration=1.859688272 podStartE2EDuration="2.416428875s" podCreationTimestamp="2025-12-06 11:01:21 +0000 UTC" firstStartedPulling="2025-12-06 11:01:22.407159955 +0000 UTC m=+1487.250591404" lastFinishedPulling="2025-12-06 11:01:22.963900568 +0000 UTC m=+1487.807332007" observedRunningTime="2025-12-06 11:01:23.404963492 +0000 UTC m=+1488.248394931" watchObservedRunningTime="2025-12-06 11:01:23.416428875 +0000 UTC m=+1488.259860314" Dec 06 11:01:44 crc kubenswrapper[4678]: I1206 11:01:44.507935 4678 scope.go:117] "RemoveContainer" containerID="58709e930d1cfb066f66ad4f8a5d037d014cd421a4650834889f677ea07f4b15" Dec 06 11:01:44 crc kubenswrapper[4678]: I1206 11:01:44.541056 4678 scope.go:117] "RemoveContainer" containerID="318c3bef97a83b0cbc09c87af9eff979f07e306aa5f9bd07e6ca3aad5c757815" Dec 06 11:01:44 crc kubenswrapper[4678]: I1206 11:01:44.585477 4678 scope.go:117] "RemoveContainer" containerID="130cf8c56956d6e41ce3133a71346ec38be43b34389acfd82e645c06b8e6e1d0" Dec 06 11:01:59 crc kubenswrapper[4678]: I1206 11:01:59.505307 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:01:59 crc kubenswrapper[4678]: I1206 11:01:59.505951 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:02:29 crc kubenswrapper[4678]: I1206 11:02:29.505422 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:02:29 crc kubenswrapper[4678]: I1206 11:02:29.506240 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.094396 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hdv7k"] Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.096723 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.107557 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hdv7k"] Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.229108 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-utilities\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.229168 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-catalog-content\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.229220 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpwfl\" (UniqueName: \"kubernetes.io/projected/d1f96431-89d9-4b57-beb1-cdc5da305400-kube-api-access-hpwfl\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.332728 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-utilities\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.333080 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-catalog-content\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.333267 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpwfl\" (UniqueName: \"kubernetes.io/projected/d1f96431-89d9-4b57-beb1-cdc5da305400-kube-api-access-hpwfl\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.333419 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-catalog-content\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.333337 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-utilities\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.356323 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpwfl\" (UniqueName: \"kubernetes.io/projected/d1f96431-89d9-4b57-beb1-cdc5da305400-kube-api-access-hpwfl\") pod \"redhat-marketplace-hdv7k\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.413915 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:36 crc kubenswrapper[4678]: I1206 11:02:36.871948 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hdv7k"] Dec 06 11:02:37 crc kubenswrapper[4678]: I1206 11:02:37.188435 4678 generic.go:334] "Generic (PLEG): container finished" podID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerID="f5dedb8605288de355a8c5985e6316ff3a4a5e66a0ad788424e803cd26e0b01d" exitCode=0 Dec 06 11:02:37 crc kubenswrapper[4678]: I1206 11:02:37.188482 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hdv7k" event={"ID":"d1f96431-89d9-4b57-beb1-cdc5da305400","Type":"ContainerDied","Data":"f5dedb8605288de355a8c5985e6316ff3a4a5e66a0ad788424e803cd26e0b01d"} Dec 06 11:02:37 crc kubenswrapper[4678]: I1206 11:02:37.188521 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hdv7k" event={"ID":"d1f96431-89d9-4b57-beb1-cdc5da305400","Type":"ContainerStarted","Data":"931323de01057a08dc81d748594e73c4ffb853a64ec441b4ef20f38d4bf6c9d9"} Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.208607 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hdv7k" event={"ID":"d1f96431-89d9-4b57-beb1-cdc5da305400","Type":"ContainerStarted","Data":"cd9f48ba61fb865dceb1d3b0e300a2f28ad2bb7a553679d3b216a45aaee01e08"} Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.667824 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wdvkn"] Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.670872 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.702472 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdvkn"] Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.804055 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-catalog-content\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.804126 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-utilities\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.804393 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd8wq\" (UniqueName: \"kubernetes.io/projected/24f1e9cc-db7d-489d-99c5-46e2d5767b49-kube-api-access-fd8wq\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.906528 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-catalog-content\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.906768 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-utilities\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.907018 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd8wq\" (UniqueName: \"kubernetes.io/projected/24f1e9cc-db7d-489d-99c5-46e2d5767b49-kube-api-access-fd8wq\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.907086 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-catalog-content\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.907617 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-utilities\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:39 crc kubenswrapper[4678]: I1206 11:02:39.929399 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd8wq\" (UniqueName: \"kubernetes.io/projected/24f1e9cc-db7d-489d-99c5-46e2d5767b49-kube-api-access-fd8wq\") pod \"certified-operators-wdvkn\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:40 crc kubenswrapper[4678]: I1206 11:02:40.035221 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:40 crc kubenswrapper[4678]: I1206 11:02:40.232528 4678 generic.go:334] "Generic (PLEG): container finished" podID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerID="cd9f48ba61fb865dceb1d3b0e300a2f28ad2bb7a553679d3b216a45aaee01e08" exitCode=0 Dec 06 11:02:40 crc kubenswrapper[4678]: I1206 11:02:40.232714 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hdv7k" event={"ID":"d1f96431-89d9-4b57-beb1-cdc5da305400","Type":"ContainerDied","Data":"cd9f48ba61fb865dceb1d3b0e300a2f28ad2bb7a553679d3b216a45aaee01e08"} Dec 06 11:02:40 crc kubenswrapper[4678]: I1206 11:02:40.551183 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wdvkn"] Dec 06 11:02:41 crc kubenswrapper[4678]: I1206 11:02:41.243783 4678 generic.go:334] "Generic (PLEG): container finished" podID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerID="686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af" exitCode=0 Dec 06 11:02:41 crc kubenswrapper[4678]: I1206 11:02:41.243882 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdvkn" event={"ID":"24f1e9cc-db7d-489d-99c5-46e2d5767b49","Type":"ContainerDied","Data":"686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af"} Dec 06 11:02:41 crc kubenswrapper[4678]: I1206 11:02:41.244169 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdvkn" event={"ID":"24f1e9cc-db7d-489d-99c5-46e2d5767b49","Type":"ContainerStarted","Data":"c0df4acd858416898b1cbd9cfaf1ab1c05ca6fc6a0b770a3797105f4d2dfc84a"} Dec 06 11:02:41 crc kubenswrapper[4678]: I1206 11:02:41.249548 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hdv7k" event={"ID":"d1f96431-89d9-4b57-beb1-cdc5da305400","Type":"ContainerStarted","Data":"74c7cee84aa2b1df3deb15ba7c039b13fa352a76d0712b6dda21236564b9dced"} Dec 06 11:02:41 crc kubenswrapper[4678]: I1206 11:02:41.296867 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hdv7k" podStartSLOduration=1.491985828 podStartE2EDuration="5.296851583s" podCreationTimestamp="2025-12-06 11:02:36 +0000 UTC" firstStartedPulling="2025-12-06 11:02:37.190044454 +0000 UTC m=+1562.033475893" lastFinishedPulling="2025-12-06 11:02:40.994910209 +0000 UTC m=+1565.838341648" observedRunningTime="2025-12-06 11:02:41.288640976 +0000 UTC m=+1566.132072435" watchObservedRunningTime="2025-12-06 11:02:41.296851583 +0000 UTC m=+1566.140283022" Dec 06 11:02:42 crc kubenswrapper[4678]: I1206 11:02:42.263609 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdvkn" event={"ID":"24f1e9cc-db7d-489d-99c5-46e2d5767b49","Type":"ContainerStarted","Data":"69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74"} Dec 06 11:02:44 crc kubenswrapper[4678]: I1206 11:02:44.280008 4678 generic.go:334] "Generic (PLEG): container finished" podID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerID="69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74" exitCode=0 Dec 06 11:02:44 crc kubenswrapper[4678]: I1206 11:02:44.280087 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdvkn" event={"ID":"24f1e9cc-db7d-489d-99c5-46e2d5767b49","Type":"ContainerDied","Data":"69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74"} Dec 06 11:02:44 crc kubenswrapper[4678]: I1206 11:02:44.779129 4678 scope.go:117] "RemoveContainer" containerID="1cc7475c58e76a748f418936dc0f0af1b64deac6832d1a31e3f736f45e4011cc" Dec 06 11:02:44 crc kubenswrapper[4678]: I1206 11:02:44.825663 4678 scope.go:117] "RemoveContainer" containerID="0c6771cf0d5d25a0f33802927088ac645141b77ff4375191cdcd0be341faeea3" Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.018509 4678 scope.go:117] "RemoveContainer" containerID="5a73f5df43ec07225878944207d903870895424133964a5124b8e351950f50fa" Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.036607 4678 scope.go:117] "RemoveContainer" containerID="0964ceddf233314aea34efa77bab3f1761d93ba02156121900794bb4df1ede45" Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.058934 4678 scope.go:117] "RemoveContainer" containerID="f20d4dcd8b390c979698d3d4e59545b67f8a42853778bd20da961122a56f8010" Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.282433 4678 scope.go:117] "RemoveContainer" containerID="6200b7dd3e61dde1e396ff21451edde208c22b656f3b16792eddc1031dc13ecb" Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.306996 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdvkn" event={"ID":"24f1e9cc-db7d-489d-99c5-46e2d5767b49","Type":"ContainerStarted","Data":"7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb"} Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.320481 4678 scope.go:117] "RemoveContainer" containerID="a0e6f9543fba06b3c541e2257f838486ec66f360a9c1271df1d580d070760d62" Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.338782 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wdvkn" podStartSLOduration=2.922287758 podStartE2EDuration="6.338759542s" podCreationTimestamp="2025-12-06 11:02:39 +0000 UTC" firstStartedPulling="2025-12-06 11:02:41.246784465 +0000 UTC m=+1566.090215904" lastFinishedPulling="2025-12-06 11:02:44.663256239 +0000 UTC m=+1569.506687688" observedRunningTime="2025-12-06 11:02:45.334132731 +0000 UTC m=+1570.177564170" watchObservedRunningTime="2025-12-06 11:02:45.338759542 +0000 UTC m=+1570.182190981" Dec 06 11:02:45 crc kubenswrapper[4678]: I1206 11:02:45.382017 4678 scope.go:117] "RemoveContainer" containerID="b0facf2a6703c559ab4d6fd9a8fe36d4168c234d46da2104728642e1cc47fff1" Dec 06 11:02:46 crc kubenswrapper[4678]: I1206 11:02:46.415206 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:46 crc kubenswrapper[4678]: I1206 11:02:46.415659 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:46 crc kubenswrapper[4678]: I1206 11:02:46.469433 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:47 crc kubenswrapper[4678]: I1206 11:02:47.383839 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:47 crc kubenswrapper[4678]: I1206 11:02:47.658179 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hdv7k"] Dec 06 11:02:49 crc kubenswrapper[4678]: I1206 11:02:49.348120 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hdv7k" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="registry-server" containerID="cri-o://74c7cee84aa2b1df3deb15ba7c039b13fa352a76d0712b6dda21236564b9dced" gracePeriod=2 Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.036027 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.036268 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.093164 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.361359 4678 generic.go:334] "Generic (PLEG): container finished" podID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerID="74c7cee84aa2b1df3deb15ba7c039b13fa352a76d0712b6dda21236564b9dced" exitCode=0 Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.362649 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hdv7k" event={"ID":"d1f96431-89d9-4b57-beb1-cdc5da305400","Type":"ContainerDied","Data":"74c7cee84aa2b1df3deb15ba7c039b13fa352a76d0712b6dda21236564b9dced"} Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.362686 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hdv7k" event={"ID":"d1f96431-89d9-4b57-beb1-cdc5da305400","Type":"ContainerDied","Data":"931323de01057a08dc81d748594e73c4ffb853a64ec441b4ef20f38d4bf6c9d9"} Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.362702 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="931323de01057a08dc81d748594e73c4ffb853a64ec441b4ef20f38d4bf6c9d9" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.393149 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.411462 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.518397 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-catalog-content\") pod \"d1f96431-89d9-4b57-beb1-cdc5da305400\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.518544 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-utilities\") pod \"d1f96431-89d9-4b57-beb1-cdc5da305400\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.518576 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpwfl\" (UniqueName: \"kubernetes.io/projected/d1f96431-89d9-4b57-beb1-cdc5da305400-kube-api-access-hpwfl\") pod \"d1f96431-89d9-4b57-beb1-cdc5da305400\" (UID: \"d1f96431-89d9-4b57-beb1-cdc5da305400\") " Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.519225 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-utilities" (OuterVolumeSpecName: "utilities") pod "d1f96431-89d9-4b57-beb1-cdc5da305400" (UID: "d1f96431-89d9-4b57-beb1-cdc5da305400"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.539111 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1f96431-89d9-4b57-beb1-cdc5da305400" (UID: "d1f96431-89d9-4b57-beb1-cdc5da305400"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.540161 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f96431-89d9-4b57-beb1-cdc5da305400-kube-api-access-hpwfl" (OuterVolumeSpecName: "kube-api-access-hpwfl") pod "d1f96431-89d9-4b57-beb1-cdc5da305400" (UID: "d1f96431-89d9-4b57-beb1-cdc5da305400"). InnerVolumeSpecName "kube-api-access-hpwfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.621179 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.621210 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1f96431-89d9-4b57-beb1-cdc5da305400-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:02:50 crc kubenswrapper[4678]: I1206 11:02:50.621225 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpwfl\" (UniqueName: \"kubernetes.io/projected/d1f96431-89d9-4b57-beb1-cdc5da305400-kube-api-access-hpwfl\") on node \"crc\" DevicePath \"\"" Dec 06 11:02:51 crc kubenswrapper[4678]: I1206 11:02:51.368949 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hdv7k" Dec 06 11:02:51 crc kubenswrapper[4678]: I1206 11:02:51.400779 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hdv7k"] Dec 06 11:02:51 crc kubenswrapper[4678]: I1206 11:02:51.413192 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hdv7k"] Dec 06 11:02:51 crc kubenswrapper[4678]: I1206 11:02:51.485272 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" path="/var/lib/kubelet/pods/d1f96431-89d9-4b57-beb1-cdc5da305400/volumes" Dec 06 11:02:52 crc kubenswrapper[4678]: I1206 11:02:52.061394 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdvkn"] Dec 06 11:02:52 crc kubenswrapper[4678]: I1206 11:02:52.377696 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wdvkn" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="registry-server" containerID="cri-o://7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb" gracePeriod=2 Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.048769 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.169672 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-utilities\") pod \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.169850 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd8wq\" (UniqueName: \"kubernetes.io/projected/24f1e9cc-db7d-489d-99c5-46e2d5767b49-kube-api-access-fd8wq\") pod \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.169889 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-catalog-content\") pod \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\" (UID: \"24f1e9cc-db7d-489d-99c5-46e2d5767b49\") " Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.170613 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-utilities" (OuterVolumeSpecName: "utilities") pod "24f1e9cc-db7d-489d-99c5-46e2d5767b49" (UID: "24f1e9cc-db7d-489d-99c5-46e2d5767b49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.174906 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24f1e9cc-db7d-489d-99c5-46e2d5767b49-kube-api-access-fd8wq" (OuterVolumeSpecName: "kube-api-access-fd8wq") pod "24f1e9cc-db7d-489d-99c5-46e2d5767b49" (UID: "24f1e9cc-db7d-489d-99c5-46e2d5767b49"). InnerVolumeSpecName "kube-api-access-fd8wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.219329 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24f1e9cc-db7d-489d-99c5-46e2d5767b49" (UID: "24f1e9cc-db7d-489d-99c5-46e2d5767b49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.271888 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd8wq\" (UniqueName: \"kubernetes.io/projected/24f1e9cc-db7d-489d-99c5-46e2d5767b49-kube-api-access-fd8wq\") on node \"crc\" DevicePath \"\"" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.271931 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.271941 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24f1e9cc-db7d-489d-99c5-46e2d5767b49-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.388995 4678 generic.go:334] "Generic (PLEG): container finished" podID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerID="7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb" exitCode=0 Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.389042 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdvkn" event={"ID":"24f1e9cc-db7d-489d-99c5-46e2d5767b49","Type":"ContainerDied","Data":"7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb"} Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.389079 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wdvkn" event={"ID":"24f1e9cc-db7d-489d-99c5-46e2d5767b49","Type":"ContainerDied","Data":"c0df4acd858416898b1cbd9cfaf1ab1c05ca6fc6a0b770a3797105f4d2dfc84a"} Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.389087 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wdvkn" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.389118 4678 scope.go:117] "RemoveContainer" containerID="7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.423117 4678 scope.go:117] "RemoveContainer" containerID="69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.427296 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wdvkn"] Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.440289 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wdvkn"] Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.466670 4678 scope.go:117] "RemoveContainer" containerID="686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.502107 4678 scope.go:117] "RemoveContainer" containerID="7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.504367 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" path="/var/lib/kubelet/pods/24f1e9cc-db7d-489d-99c5-46e2d5767b49/volumes" Dec 06 11:02:53 crc kubenswrapper[4678]: E1206 11:02:53.505406 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb\": container with ID starting with 7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb not found: ID does not exist" containerID="7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.505436 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb"} err="failed to get container status \"7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb\": rpc error: code = NotFound desc = could not find container \"7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb\": container with ID starting with 7a835832079c553ab4e304003c83b2bb4e61fcbf998f94f332be44b999375efb not found: ID does not exist" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.505465 4678 scope.go:117] "RemoveContainer" containerID="69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74" Dec 06 11:02:53 crc kubenswrapper[4678]: E1206 11:02:53.505750 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74\": container with ID starting with 69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74 not found: ID does not exist" containerID="69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.505780 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74"} err="failed to get container status \"69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74\": rpc error: code = NotFound desc = could not find container \"69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74\": container with ID starting with 69dcc557411c6f3c4203e6c2cb8893bf98b99625980a1f356937ea580ccb6d74 not found: ID does not exist" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.505798 4678 scope.go:117] "RemoveContainer" containerID="686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af" Dec 06 11:02:53 crc kubenswrapper[4678]: E1206 11:02:53.506010 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af\": container with ID starting with 686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af not found: ID does not exist" containerID="686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af" Dec 06 11:02:53 crc kubenswrapper[4678]: I1206 11:02:53.506033 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af"} err="failed to get container status \"686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af\": rpc error: code = NotFound desc = could not find container \"686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af\": container with ID starting with 686737fea0a535e200fd0a006073d2d4d59993b60e19d0a2ba11f36b9b7363af not found: ID does not exist" Dec 06 11:02:59 crc kubenswrapper[4678]: I1206 11:02:59.505848 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:02:59 crc kubenswrapper[4678]: I1206 11:02:59.506424 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:02:59 crc kubenswrapper[4678]: I1206 11:02:59.506837 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:02:59 crc kubenswrapper[4678]: I1206 11:02:59.508193 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:02:59 crc kubenswrapper[4678]: I1206 11:02:59.508313 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" gracePeriod=600 Dec 06 11:02:59 crc kubenswrapper[4678]: E1206 11:02:59.640780 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:03:00 crc kubenswrapper[4678]: I1206 11:03:00.487674 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" exitCode=0 Dec 06 11:03:00 crc kubenswrapper[4678]: I1206 11:03:00.487720 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da"} Dec 06 11:03:00 crc kubenswrapper[4678]: I1206 11:03:00.487776 4678 scope.go:117] "RemoveContainer" containerID="c7aaeb71fc254443c399a6d76a377d0a6c35a168b9037ac2247e9eecdf49d005" Dec 06 11:03:00 crc kubenswrapper[4678]: I1206 11:03:00.488973 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:03:00 crc kubenswrapper[4678]: E1206 11:03:00.492999 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.816086 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k569s"] Dec 06 11:03:05 crc kubenswrapper[4678]: E1206 11:03:05.817094 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="registry-server" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817110 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="registry-server" Dec 06 11:03:05 crc kubenswrapper[4678]: E1206 11:03:05.817127 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="extract-content" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817135 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="extract-content" Dec 06 11:03:05 crc kubenswrapper[4678]: E1206 11:03:05.817146 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="extract-content" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817152 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="extract-content" Dec 06 11:03:05 crc kubenswrapper[4678]: E1206 11:03:05.817167 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="extract-utilities" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817175 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="extract-utilities" Dec 06 11:03:05 crc kubenswrapper[4678]: E1206 11:03:05.817199 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="registry-server" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817207 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="registry-server" Dec 06 11:03:05 crc kubenswrapper[4678]: E1206 11:03:05.817237 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="extract-utilities" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817245 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="extract-utilities" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817472 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="24f1e9cc-db7d-489d-99c5-46e2d5767b49" containerName="registry-server" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.817567 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f96431-89d9-4b57-beb1-cdc5da305400" containerName="registry-server" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.819239 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.834427 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k569s"] Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.967864 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v7fw\" (UniqueName: \"kubernetes.io/projected/7430e612-4d8a-4d48-9d55-0640d2133e06-kube-api-access-7v7fw\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.967941 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-catalog-content\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:05 crc kubenswrapper[4678]: I1206 11:03:05.967980 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-utilities\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.070027 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v7fw\" (UniqueName: \"kubernetes.io/projected/7430e612-4d8a-4d48-9d55-0640d2133e06-kube-api-access-7v7fw\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.070101 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-catalog-content\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.070143 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-utilities\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.070586 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-utilities\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.070878 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-catalog-content\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.087838 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v7fw\" (UniqueName: \"kubernetes.io/projected/7430e612-4d8a-4d48-9d55-0640d2133e06-kube-api-access-7v7fw\") pod \"community-operators-k569s\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.141676 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:06 crc kubenswrapper[4678]: I1206 11:03:06.658507 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k569s"] Dec 06 11:03:07 crc kubenswrapper[4678]: I1206 11:03:07.568084 4678 generic.go:334] "Generic (PLEG): container finished" podID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerID="bba15e5a85827ed29d73a65cf5b59e811816c1a850b6b7be87391216dd44b09e" exitCode=0 Dec 06 11:03:07 crc kubenswrapper[4678]: I1206 11:03:07.568195 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k569s" event={"ID":"7430e612-4d8a-4d48-9d55-0640d2133e06","Type":"ContainerDied","Data":"bba15e5a85827ed29d73a65cf5b59e811816c1a850b6b7be87391216dd44b09e"} Dec 06 11:03:07 crc kubenswrapper[4678]: I1206 11:03:07.568436 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k569s" event={"ID":"7430e612-4d8a-4d48-9d55-0640d2133e06","Type":"ContainerStarted","Data":"3ea786e51a9a54ffeaca508f1270b90e616c82fff2d19411f77d5f11bd4ab666"} Dec 06 11:03:09 crc kubenswrapper[4678]: I1206 11:03:09.606421 4678 generic.go:334] "Generic (PLEG): container finished" podID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerID="84ad7e0a8680bb4a94472668acfe78f03f719f4af041c6810450fa75cb821b7e" exitCode=0 Dec 06 11:03:09 crc kubenswrapper[4678]: I1206 11:03:09.606530 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k569s" event={"ID":"7430e612-4d8a-4d48-9d55-0640d2133e06","Type":"ContainerDied","Data":"84ad7e0a8680bb4a94472668acfe78f03f719f4af041c6810450fa75cb821b7e"} Dec 06 11:03:10 crc kubenswrapper[4678]: I1206 11:03:10.620786 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k569s" event={"ID":"7430e612-4d8a-4d48-9d55-0640d2133e06","Type":"ContainerStarted","Data":"3e82e55107a4289d587b66db275514cc5c0642e231350fa4aeae117e47b900ab"} Dec 06 11:03:10 crc kubenswrapper[4678]: I1206 11:03:10.643851 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k569s" podStartSLOduration=3.152119133 podStartE2EDuration="5.643833059s" podCreationTimestamp="2025-12-06 11:03:05 +0000 UTC" firstStartedPulling="2025-12-06 11:03:07.569657145 +0000 UTC m=+1592.413088584" lastFinishedPulling="2025-12-06 11:03:10.061371071 +0000 UTC m=+1594.904802510" observedRunningTime="2025-12-06 11:03:10.638985007 +0000 UTC m=+1595.482416466" watchObservedRunningTime="2025-12-06 11:03:10.643833059 +0000 UTC m=+1595.487264508" Dec 06 11:03:12 crc kubenswrapper[4678]: I1206 11:03:12.476342 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:03:12 crc kubenswrapper[4678]: E1206 11:03:12.476747 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:03:16 crc kubenswrapper[4678]: I1206 11:03:16.141850 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:16 crc kubenswrapper[4678]: I1206 11:03:16.142353 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:16 crc kubenswrapper[4678]: I1206 11:03:16.185310 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:16 crc kubenswrapper[4678]: I1206 11:03:16.802213 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:16 crc kubenswrapper[4678]: I1206 11:03:16.867310 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k569s"] Dec 06 11:03:18 crc kubenswrapper[4678]: I1206 11:03:18.720550 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k569s" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="registry-server" containerID="cri-o://3e82e55107a4289d587b66db275514cc5c0642e231350fa4aeae117e47b900ab" gracePeriod=2 Dec 06 11:03:18 crc kubenswrapper[4678]: E1206 11:03:18.803545 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7430e612_4d8a_4d48_9d55_0640d2133e06.slice/crio-3e82e55107a4289d587b66db275514cc5c0642e231350fa4aeae117e47b900ab.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:03:19 crc kubenswrapper[4678]: I1206 11:03:19.733576 4678 generic.go:334] "Generic (PLEG): container finished" podID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerID="3e82e55107a4289d587b66db275514cc5c0642e231350fa4aeae117e47b900ab" exitCode=0 Dec 06 11:03:19 crc kubenswrapper[4678]: I1206 11:03:19.733647 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k569s" event={"ID":"7430e612-4d8a-4d48-9d55-0640d2133e06","Type":"ContainerDied","Data":"3e82e55107a4289d587b66db275514cc5c0642e231350fa4aeae117e47b900ab"} Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.360089 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.446097 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-catalog-content\") pod \"7430e612-4d8a-4d48-9d55-0640d2133e06\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.446170 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-utilities\") pod \"7430e612-4d8a-4d48-9d55-0640d2133e06\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.446253 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v7fw\" (UniqueName: \"kubernetes.io/projected/7430e612-4d8a-4d48-9d55-0640d2133e06-kube-api-access-7v7fw\") pod \"7430e612-4d8a-4d48-9d55-0640d2133e06\" (UID: \"7430e612-4d8a-4d48-9d55-0640d2133e06\") " Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.447095 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-utilities" (OuterVolumeSpecName: "utilities") pod "7430e612-4d8a-4d48-9d55-0640d2133e06" (UID: "7430e612-4d8a-4d48-9d55-0640d2133e06"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.448196 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.470134 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7430e612-4d8a-4d48-9d55-0640d2133e06-kube-api-access-7v7fw" (OuterVolumeSpecName: "kube-api-access-7v7fw") pod "7430e612-4d8a-4d48-9d55-0640d2133e06" (UID: "7430e612-4d8a-4d48-9d55-0640d2133e06"). InnerVolumeSpecName "kube-api-access-7v7fw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.501113 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7430e612-4d8a-4d48-9d55-0640d2133e06" (UID: "7430e612-4d8a-4d48-9d55-0640d2133e06"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.549932 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7430e612-4d8a-4d48-9d55-0640d2133e06-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.549973 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v7fw\" (UniqueName: \"kubernetes.io/projected/7430e612-4d8a-4d48-9d55-0640d2133e06-kube-api-access-7v7fw\") on node \"crc\" DevicePath \"\"" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.743939 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k569s" event={"ID":"7430e612-4d8a-4d48-9d55-0640d2133e06","Type":"ContainerDied","Data":"3ea786e51a9a54ffeaca508f1270b90e616c82fff2d19411f77d5f11bd4ab666"} Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.744253 4678 scope.go:117] "RemoveContainer" containerID="3e82e55107a4289d587b66db275514cc5c0642e231350fa4aeae117e47b900ab" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.744211 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k569s" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.775545 4678 scope.go:117] "RemoveContainer" containerID="84ad7e0a8680bb4a94472668acfe78f03f719f4af041c6810450fa75cb821b7e" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.791733 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k569s"] Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.805965 4678 scope.go:117] "RemoveContainer" containerID="bba15e5a85827ed29d73a65cf5b59e811816c1a850b6b7be87391216dd44b09e" Dec 06 11:03:20 crc kubenswrapper[4678]: I1206 11:03:20.810565 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k569s"] Dec 06 11:03:21 crc kubenswrapper[4678]: I1206 11:03:21.494170 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" path="/var/lib/kubelet/pods/7430e612-4d8a-4d48-9d55-0640d2133e06/volumes" Dec 06 11:03:25 crc kubenswrapper[4678]: I1206 11:03:25.485768 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:03:25 crc kubenswrapper[4678]: E1206 11:03:25.486701 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:03:39 crc kubenswrapper[4678]: I1206 11:03:39.476009 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:03:39 crc kubenswrapper[4678]: E1206 11:03:39.476819 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:03:45 crc kubenswrapper[4678]: I1206 11:03:45.495919 4678 scope.go:117] "RemoveContainer" containerID="d0eb0137b8e92d5cf5437f09289085a181f428fa4b015276e1e04690908dc91c" Dec 06 11:03:50 crc kubenswrapper[4678]: I1206 11:03:50.476568 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:03:50 crc kubenswrapper[4678]: E1206 11:03:50.477380 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:04:01 crc kubenswrapper[4678]: I1206 11:04:01.477023 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:04:01 crc kubenswrapper[4678]: E1206 11:04:01.478003 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:04:16 crc kubenswrapper[4678]: I1206 11:04:16.476805 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:04:16 crc kubenswrapper[4678]: E1206 11:04:16.477553 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:04:30 crc kubenswrapper[4678]: I1206 11:04:30.475955 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:04:30 crc kubenswrapper[4678]: E1206 11:04:30.476712 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:04:37 crc kubenswrapper[4678]: I1206 11:04:37.053305 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-mwspd"] Dec 06 11:04:37 crc kubenswrapper[4678]: I1206 11:04:37.067378 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-mwspd"] Dec 06 11:04:37 crc kubenswrapper[4678]: I1206 11:04:37.489344 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b" path="/var/lib/kubelet/pods/e7b3e1f0-ed17-41e7-a4b6-60f8a247af2b/volumes" Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.036022 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-98b3-account-create-update-4rg7b"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.051139 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-5mckr"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.065848 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-xvjr7"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.075024 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-e1f4-account-create-update-pnscz"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.084499 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-17f0-account-create-update-8xfqp"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.095688 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-98b3-account-create-update-4rg7b"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.106077 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-5mckr"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.114710 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-xvjr7"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.123518 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-e1f4-account-create-update-pnscz"] Dec 06 11:04:38 crc kubenswrapper[4678]: I1206 11:04:38.131478 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-17f0-account-create-update-8xfqp"] Dec 06 11:04:39 crc kubenswrapper[4678]: I1206 11:04:39.491207 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f077be8-3497-464c-9bfd-4752c6bc08c3" path="/var/lib/kubelet/pods/2f077be8-3497-464c-9bfd-4752c6bc08c3/volumes" Dec 06 11:04:39 crc kubenswrapper[4678]: I1206 11:04:39.494068 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51f8e528-9dfd-4572-9ad6-07d8393be1bb" path="/var/lib/kubelet/pods/51f8e528-9dfd-4572-9ad6-07d8393be1bb/volumes" Dec 06 11:04:39 crc kubenswrapper[4678]: I1206 11:04:39.496248 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b226f5f5-03ea-4614-bd24-5499b6fe7457" path="/var/lib/kubelet/pods/b226f5f5-03ea-4614-bd24-5499b6fe7457/volumes" Dec 06 11:04:39 crc kubenswrapper[4678]: I1206 11:04:39.498282 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5efdf59-1cdd-4aef-8105-f44519426d2c" path="/var/lib/kubelet/pods/c5efdf59-1cdd-4aef-8105-f44519426d2c/volumes" Dec 06 11:04:39 crc kubenswrapper[4678]: I1206 11:04:39.500606 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f09ba32d-29da-4907-80b0-4df3433704f8" path="/var/lib/kubelet/pods/f09ba32d-29da-4907-80b0-4df3433704f8/volumes" Dec 06 11:04:44 crc kubenswrapper[4678]: I1206 11:04:44.480456 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:04:44 crc kubenswrapper[4678]: E1206 11:04:44.480898 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:04:45 crc kubenswrapper[4678]: I1206 11:04:45.603261 4678 scope.go:117] "RemoveContainer" containerID="5c1b064acf9cadbd8425c6c46461651efbbe9097287298ae7d003d9ce0411c33" Dec 06 11:04:45 crc kubenswrapper[4678]: I1206 11:04:45.644961 4678 scope.go:117] "RemoveContainer" containerID="b0a5cbee3899208a14d1322cbf9a30eb3f16965598f6873e3e7e9e9195d57916" Dec 06 11:04:45 crc kubenswrapper[4678]: I1206 11:04:45.689560 4678 scope.go:117] "RemoveContainer" containerID="e5984f9e8537e9c8f3f4248f254aed3cad0f0398bce3fb3abfea5471cda418e5" Dec 06 11:04:45 crc kubenswrapper[4678]: I1206 11:04:45.755242 4678 scope.go:117] "RemoveContainer" containerID="60408b295cd3659ac8b9add1dcd546a11edddc50c9450b186c1a2d352c6434e3" Dec 06 11:04:45 crc kubenswrapper[4678]: I1206 11:04:45.790613 4678 scope.go:117] "RemoveContainer" containerID="2d4260b04317a574b7d3e38bee42021cf319114d7ce64e95ac732240971fe18c" Dec 06 11:04:45 crc kubenswrapper[4678]: I1206 11:04:45.848149 4678 scope.go:117] "RemoveContainer" containerID="05d55e96ccf43e190ec08bef298678e3ac341df89d7b5503b74c1dd34bf249be" Dec 06 11:04:48 crc kubenswrapper[4678]: I1206 11:04:48.675365 4678 generic.go:334] "Generic (PLEG): container finished" podID="99d806d5-73ca-4ecc-92cb-b3c300cb04b4" containerID="f6ab2f1feebb265a01d5f45c9e965958c1f768f78e9e7914c6b9d3f3fbc01208" exitCode=0 Dec 06 11:04:48 crc kubenswrapper[4678]: I1206 11:04:48.675454 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" event={"ID":"99d806d5-73ca-4ecc-92cb-b3c300cb04b4","Type":"ContainerDied","Data":"f6ab2f1feebb265a01d5f45c9e965958c1f768f78e9e7914c6b9d3f3fbc01208"} Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.170979 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.284079 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-ssh-key\") pod \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.284125 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6c7b\" (UniqueName: \"kubernetes.io/projected/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-kube-api-access-q6c7b\") pod \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.284241 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-inventory\") pod \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.284291 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-bootstrap-combined-ca-bundle\") pod \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\" (UID: \"99d806d5-73ca-4ecc-92cb-b3c300cb04b4\") " Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.296522 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "99d806d5-73ca-4ecc-92cb-b3c300cb04b4" (UID: "99d806d5-73ca-4ecc-92cb-b3c300cb04b4"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.297686 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-kube-api-access-q6c7b" (OuterVolumeSpecName: "kube-api-access-q6c7b") pod "99d806d5-73ca-4ecc-92cb-b3c300cb04b4" (UID: "99d806d5-73ca-4ecc-92cb-b3c300cb04b4"). InnerVolumeSpecName "kube-api-access-q6c7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.310431 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-inventory" (OuterVolumeSpecName: "inventory") pod "99d806d5-73ca-4ecc-92cb-b3c300cb04b4" (UID: "99d806d5-73ca-4ecc-92cb-b3c300cb04b4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.310911 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99d806d5-73ca-4ecc-92cb-b3c300cb04b4" (UID: "99d806d5-73ca-4ecc-92cb-b3c300cb04b4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.386305 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.386569 4678 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.386639 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.386736 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6c7b\" (UniqueName: \"kubernetes.io/projected/99d806d5-73ca-4ecc-92cb-b3c300cb04b4-kube-api-access-q6c7b\") on node \"crc\" DevicePath \"\"" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.692322 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" event={"ID":"99d806d5-73ca-4ecc-92cb-b3c300cb04b4","Type":"ContainerDied","Data":"47cc28df0829e7ded6349f3dfa8b4fd1d57c93fcf998e1d842a105178bfb8039"} Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.692394 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47cc28df0829e7ded6349f3dfa8b4fd1d57c93fcf998e1d842a105178bfb8039" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.692397 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.787004 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb"] Dec 06 11:04:50 crc kubenswrapper[4678]: E1206 11:04:50.787704 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="registry-server" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.787792 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="registry-server" Dec 06 11:04:50 crc kubenswrapper[4678]: E1206 11:04:50.787901 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="extract-utilities" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.787955 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="extract-utilities" Dec 06 11:04:50 crc kubenswrapper[4678]: E1206 11:04:50.788029 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d806d5-73ca-4ecc-92cb-b3c300cb04b4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.788091 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d806d5-73ca-4ecc-92cb-b3c300cb04b4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 11:04:50 crc kubenswrapper[4678]: E1206 11:04:50.788176 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="extract-content" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.788237 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="extract-content" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.788480 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d806d5-73ca-4ecc-92cb-b3c300cb04b4" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.788564 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="7430e612-4d8a-4d48-9d55-0640d2133e06" containerName="registry-server" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.789211 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.791945 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.792135 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.792256 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.795860 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.809888 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb"] Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.896834 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.896924 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56vxk\" (UniqueName: \"kubernetes.io/projected/9ac6052e-a10a-4709-b1be-99bb37d0922d-kube-api-access-56vxk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.897149 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.999336 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.999474 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:50 crc kubenswrapper[4678]: I1206 11:04:50.999545 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56vxk\" (UniqueName: \"kubernetes.io/projected/9ac6052e-a10a-4709-b1be-99bb37d0922d-kube-api-access-56vxk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:51 crc kubenswrapper[4678]: I1206 11:04:51.003872 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:51 crc kubenswrapper[4678]: I1206 11:04:51.005292 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:51 crc kubenswrapper[4678]: I1206 11:04:51.018008 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56vxk\" (UniqueName: \"kubernetes.io/projected/9ac6052e-a10a-4709-b1be-99bb37d0922d-kube-api-access-56vxk\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-877bb\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:51 crc kubenswrapper[4678]: I1206 11:04:51.103734 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:04:51 crc kubenswrapper[4678]: I1206 11:04:51.673009 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb"] Dec 06 11:04:51 crc kubenswrapper[4678]: I1206 11:04:51.679327 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:04:51 crc kubenswrapper[4678]: I1206 11:04:51.705001 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" event={"ID":"9ac6052e-a10a-4709-b1be-99bb37d0922d","Type":"ContainerStarted","Data":"3e4db2b9cd2674dd3912a58f995f897eb1cc9fd2edb343e4589e2ff56c586dee"} Dec 06 11:04:52 crc kubenswrapper[4678]: I1206 11:04:52.718350 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" event={"ID":"9ac6052e-a10a-4709-b1be-99bb37d0922d","Type":"ContainerStarted","Data":"48439e763aeab9187b3d4004400945b94aa26abed4b9b621707fcabf213ad398"} Dec 06 11:04:52 crc kubenswrapper[4678]: I1206 11:04:52.752873 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" podStartSLOduration=2.199030325 podStartE2EDuration="2.75284924s" podCreationTimestamp="2025-12-06 11:04:50 +0000 UTC" firstStartedPulling="2025-12-06 11:04:51.678704005 +0000 UTC m=+1696.522135444" lastFinishedPulling="2025-12-06 11:04:52.23252292 +0000 UTC m=+1697.075954359" observedRunningTime="2025-12-06 11:04:52.743392192 +0000 UTC m=+1697.586823631" watchObservedRunningTime="2025-12-06 11:04:52.75284924 +0000 UTC m=+1697.596280709" Dec 06 11:04:58 crc kubenswrapper[4678]: I1206 11:04:58.475421 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:04:58 crc kubenswrapper[4678]: E1206 11:04:58.476292 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:05:08 crc kubenswrapper[4678]: I1206 11:05:08.046507 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-h5wzb"] Dec 06 11:05:08 crc kubenswrapper[4678]: I1206 11:05:08.058735 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-h5wzb"] Dec 06 11:05:09 crc kubenswrapper[4678]: I1206 11:05:09.486157 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9024ed1c-7251-4e71-aa5e-48354a9813b7" path="/var/lib/kubelet/pods/9024ed1c-7251-4e71-aa5e-48354a9813b7/volumes" Dec 06 11:05:10 crc kubenswrapper[4678]: I1206 11:05:10.477571 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:05:10 crc kubenswrapper[4678]: E1206 11:05:10.477815 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:05:20 crc kubenswrapper[4678]: I1206 11:05:20.026379 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-055f-account-create-update-wmjpx"] Dec 06 11:05:20 crc kubenswrapper[4678]: I1206 11:05:20.048319 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-055f-account-create-update-wmjpx"] Dec 06 11:05:21 crc kubenswrapper[4678]: I1206 11:05:21.491374 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a074406-0662-4e37-8cb5-4f187d687cb8" path="/var/lib/kubelet/pods/8a074406-0662-4e37-8cb5-4f187d687cb8/volumes" Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.034269 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-0275-account-create-update-sxr85"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.044899 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-x9jhf"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.055532 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qjg97"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.068221 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f054-account-create-update-f98s6"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.082632 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-m54tp"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.090818 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qjg97"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.098332 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-m54tp"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.105562 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-x9jhf"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.112865 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-0275-account-create-update-sxr85"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.119914 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f054-account-create-update-f98s6"] Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.491398 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1" path="/var/lib/kubelet/pods/54cdc7e5-a5f1-4b88-b5a6-36ed4fdd08f1/volumes" Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.493553 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d28bdf3-353f-444d-b401-ff06978869c9" path="/var/lib/kubelet/pods/7d28bdf3-353f-444d-b401-ff06978869c9/volumes" Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.495460 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97386078-2478-4899-be49-4f8bb3019081" path="/var/lib/kubelet/pods/97386078-2478-4899-be49-4f8bb3019081/volumes" Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.497157 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db" path="/var/lib/kubelet/pods/b2d3760e-f3e5-4ae3-93a7-ff1c2f0bf4db/volumes" Dec 06 11:05:23 crc kubenswrapper[4678]: I1206 11:05:23.500251 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daddc760-3153-40b0-8e44-01e32a9d909f" path="/var/lib/kubelet/pods/daddc760-3153-40b0-8e44-01e32a9d909f/volumes" Dec 06 11:05:25 crc kubenswrapper[4678]: I1206 11:05:25.486657 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:05:25 crc kubenswrapper[4678]: E1206 11:05:25.488119 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:05:28 crc kubenswrapper[4678]: I1206 11:05:28.026020 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-mzqdh"] Dec 06 11:05:28 crc kubenswrapper[4678]: I1206 11:05:28.037055 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-mzqdh"] Dec 06 11:05:29 crc kubenswrapper[4678]: I1206 11:05:29.497676 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26bbb0e2-894e-4476-b5c8-4af78dedafd8" path="/var/lib/kubelet/pods/26bbb0e2-894e-4476-b5c8-4af78dedafd8/volumes" Dec 06 11:05:37 crc kubenswrapper[4678]: I1206 11:05:37.476309 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:05:37 crc kubenswrapper[4678]: E1206 11:05:37.477308 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.034239 4678 scope.go:117] "RemoveContainer" containerID="e46c9a15f14bc7ee694c2b6999da21360588452aa5aaf7fb51a13df03e13dce8" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.087797 4678 scope.go:117] "RemoveContainer" containerID="990ce8ae0bd650b55dd57e3a758706a4534166a91d32aa8bb76b17c751b15b51" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.124325 4678 scope.go:117] "RemoveContainer" containerID="f28de67ad744a935191f8ce3699dbfab7838ac5d52dbbeb3238b39ce433043c7" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.180010 4678 scope.go:117] "RemoveContainer" containerID="b069c241c02034d05ec4f49072ca4fdc939d62f2e0d0127aa4598e6a40a950d1" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.237288 4678 scope.go:117] "RemoveContainer" containerID="0ccf83f93468cdad88fec3e6b3dc40dc30e3dab86044fbf29afd783baf1842f2" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.285672 4678 scope.go:117] "RemoveContainer" containerID="0007edd53857b7b76507a17ecfaaf83b8926043cffd1737c98f9d930f8127c02" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.322633 4678 scope.go:117] "RemoveContainer" containerID="4e4e6974c8af5150f499a772ee857cb6878a20594d12b388de39f7fc6c0be9df" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.338980 4678 scope.go:117] "RemoveContainer" containerID="435762cba1461d3cd78d31088be4ff9c7976bc87d2b1c212357b5b1d03f7bda0" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.365162 4678 scope.go:117] "RemoveContainer" containerID="c0da59169d40ad44d3181acffa370adebc4a12b3ffad4637228e8e8fdb7e9401" Dec 06 11:05:46 crc kubenswrapper[4678]: I1206 11:05:46.395226 4678 scope.go:117] "RemoveContainer" containerID="5afb9f55b56a33c076a94c6f6e31da815df753dcc4d5bf5ef8007c7348638920" Dec 06 11:05:51 crc kubenswrapper[4678]: I1206 11:05:51.476187 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:05:51 crc kubenswrapper[4678]: E1206 11:05:51.477022 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:06:05 crc kubenswrapper[4678]: I1206 11:06:05.486545 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:06:05 crc kubenswrapper[4678]: E1206 11:06:05.487456 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:06:10 crc kubenswrapper[4678]: I1206 11:06:10.040992 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-lmpj6"] Dec 06 11:06:10 crc kubenswrapper[4678]: I1206 11:06:10.054920 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-lmpj6"] Dec 06 11:06:11 crc kubenswrapper[4678]: I1206 11:06:11.493282 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cca26a10-9756-4a74-8a33-688f36c21b46" path="/var/lib/kubelet/pods/cca26a10-9756-4a74-8a33-688f36c21b46/volumes" Dec 06 11:06:20 crc kubenswrapper[4678]: I1206 11:06:20.475783 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:06:20 crc kubenswrapper[4678]: E1206 11:06:20.476532 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:06:23 crc kubenswrapper[4678]: I1206 11:06:23.045194 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-j8h79"] Dec 06 11:06:23 crc kubenswrapper[4678]: I1206 11:06:23.064476 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-j8h79"] Dec 06 11:06:23 crc kubenswrapper[4678]: I1206 11:06:23.491840 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6584c00a-aebd-4daf-b5db-e5cb59978e86" path="/var/lib/kubelet/pods/6584c00a-aebd-4daf-b5db-e5cb59978e86/volumes" Dec 06 11:06:24 crc kubenswrapper[4678]: I1206 11:06:24.030765 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-cvhc9"] Dec 06 11:06:24 crc kubenswrapper[4678]: I1206 11:06:24.042067 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-cvhc9"] Dec 06 11:06:25 crc kubenswrapper[4678]: I1206 11:06:25.491973 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fbf7713-1e99-4a84-87d0-82803fb8d0e4" path="/var/lib/kubelet/pods/8fbf7713-1e99-4a84-87d0-82803fb8d0e4/volumes" Dec 06 11:06:35 crc kubenswrapper[4678]: I1206 11:06:35.490611 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:06:35 crc kubenswrapper[4678]: E1206 11:06:35.491646 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:06:37 crc kubenswrapper[4678]: I1206 11:06:37.768057 4678 generic.go:334] "Generic (PLEG): container finished" podID="9ac6052e-a10a-4709-b1be-99bb37d0922d" containerID="48439e763aeab9187b3d4004400945b94aa26abed4b9b621707fcabf213ad398" exitCode=0 Dec 06 11:06:37 crc kubenswrapper[4678]: I1206 11:06:37.768172 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" event={"ID":"9ac6052e-a10a-4709-b1be-99bb37d0922d","Type":"ContainerDied","Data":"48439e763aeab9187b3d4004400945b94aa26abed4b9b621707fcabf213ad398"} Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.084548 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xj2n5"] Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.101975 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xj2n5"] Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.273296 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.435797 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56vxk\" (UniqueName: \"kubernetes.io/projected/9ac6052e-a10a-4709-b1be-99bb37d0922d-kube-api-access-56vxk\") pod \"9ac6052e-a10a-4709-b1be-99bb37d0922d\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.435870 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-inventory\") pod \"9ac6052e-a10a-4709-b1be-99bb37d0922d\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.436003 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-ssh-key\") pod \"9ac6052e-a10a-4709-b1be-99bb37d0922d\" (UID: \"9ac6052e-a10a-4709-b1be-99bb37d0922d\") " Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.441945 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac6052e-a10a-4709-b1be-99bb37d0922d-kube-api-access-56vxk" (OuterVolumeSpecName: "kube-api-access-56vxk") pod "9ac6052e-a10a-4709-b1be-99bb37d0922d" (UID: "9ac6052e-a10a-4709-b1be-99bb37d0922d"). InnerVolumeSpecName "kube-api-access-56vxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.478759 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9ac6052e-a10a-4709-b1be-99bb37d0922d" (UID: "9ac6052e-a10a-4709-b1be-99bb37d0922d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.481187 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-inventory" (OuterVolumeSpecName: "inventory") pod "9ac6052e-a10a-4709-b1be-99bb37d0922d" (UID: "9ac6052e-a10a-4709-b1be-99bb37d0922d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.494896 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a95e322-b0b3-4083-81a3-187b579c53f4" path="/var/lib/kubelet/pods/3a95e322-b0b3-4083-81a3-187b579c53f4/volumes" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.537667 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56vxk\" (UniqueName: \"kubernetes.io/projected/9ac6052e-a10a-4709-b1be-99bb37d0922d-kube-api-access-56vxk\") on node \"crc\" DevicePath \"\"" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.537700 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.537709 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ac6052e-a10a-4709-b1be-99bb37d0922d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.792332 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" event={"ID":"9ac6052e-a10a-4709-b1be-99bb37d0922d","Type":"ContainerDied","Data":"3e4db2b9cd2674dd3912a58f995f897eb1cc9fd2edb343e4589e2ff56c586dee"} Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.792665 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e4db2b9cd2674dd3912a58f995f897eb1cc9fd2edb343e4589e2ff56c586dee" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.792429 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-877bb" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.894016 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj"] Dec 06 11:06:39 crc kubenswrapper[4678]: E1206 11:06:39.894420 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac6052e-a10a-4709-b1be-99bb37d0922d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.894438 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac6052e-a10a-4709-b1be-99bb37d0922d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.894631 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac6052e-a10a-4709-b1be-99bb37d0922d" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.895227 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.898738 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.898927 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.899046 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.900412 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:06:39 crc kubenswrapper[4678]: I1206 11:06:39.910500 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj"] Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.025385 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-c448f"] Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.033800 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-c448f"] Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.047344 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjdgn\" (UniqueName: \"kubernetes.io/projected/5fa31310-7e0f-431a-ae13-287fd25a926c-kube-api-access-mjdgn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.047415 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.047514 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.148634 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjdgn\" (UniqueName: \"kubernetes.io/projected/5fa31310-7e0f-431a-ae13-287fd25a926c-kube-api-access-mjdgn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.148705 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.148800 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.153446 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.160292 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.164853 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjdgn\" (UniqueName: \"kubernetes.io/projected/5fa31310-7e0f-431a-ae13-287fd25a926c-kube-api-access-mjdgn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.213658 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.794669 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj"] Dec 06 11:06:40 crc kubenswrapper[4678]: I1206 11:06:40.807173 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" event={"ID":"5fa31310-7e0f-431a-ae13-287fd25a926c","Type":"ContainerStarted","Data":"2ee4c19db692dd8f7f2cfffa8d6b640a9971610ed2f0770d65e521dc2e9a5d2a"} Dec 06 11:06:41 crc kubenswrapper[4678]: I1206 11:06:41.490485 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c315badb-6f3b-408f-b68f-b359cea62f98" path="/var/lib/kubelet/pods/c315badb-6f3b-408f-b68f-b359cea62f98/volumes" Dec 06 11:06:41 crc kubenswrapper[4678]: I1206 11:06:41.822855 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" event={"ID":"5fa31310-7e0f-431a-ae13-287fd25a926c","Type":"ContainerStarted","Data":"9bb2618347ba4c503d1285af113fc9d7d2fa61fd7c0fac4678a1e6d2fbce1d4a"} Dec 06 11:06:41 crc kubenswrapper[4678]: I1206 11:06:41.859283 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" podStartSLOduration=2.365163186 podStartE2EDuration="2.859255341s" podCreationTimestamp="2025-12-06 11:06:39 +0000 UTC" firstStartedPulling="2025-12-06 11:06:40.800711128 +0000 UTC m=+1805.644142557" lastFinishedPulling="2025-12-06 11:06:41.294803273 +0000 UTC m=+1806.138234712" observedRunningTime="2025-12-06 11:06:41.842934478 +0000 UTC m=+1806.686365917" watchObservedRunningTime="2025-12-06 11:06:41.859255341 +0000 UTC m=+1806.702686780" Dec 06 11:06:46 crc kubenswrapper[4678]: I1206 11:06:46.580662 4678 scope.go:117] "RemoveContainer" containerID="13e34f933d799fe522f0bd65b00d3c3ef282f1e61d5d13adf24ebaaf6b67e908" Dec 06 11:06:46 crc kubenswrapper[4678]: I1206 11:06:46.607285 4678 scope.go:117] "RemoveContainer" containerID="3f0590eb23cf2fba33a719d3ba3fb99b45e39f94ba69a13180224a0a50bc72e2" Dec 06 11:06:46 crc kubenswrapper[4678]: I1206 11:06:46.656242 4678 scope.go:117] "RemoveContainer" containerID="66ef8a822ff594231a59cc28f0a007e97e0d17aa5883855c56365e8ee7c84ee2" Dec 06 11:06:46 crc kubenswrapper[4678]: I1206 11:06:46.710583 4678 scope.go:117] "RemoveContainer" containerID="4f8fd36da47d9da3f108b03408507b0bf225441036bb98d6d84c8a026018f317" Dec 06 11:06:46 crc kubenswrapper[4678]: I1206 11:06:46.758452 4678 scope.go:117] "RemoveContainer" containerID="a5649c76e4a471f9fd4a47f9a86959923c310d68810204b448fc42b47cab289c" Dec 06 11:06:47 crc kubenswrapper[4678]: I1206 11:06:47.476871 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:06:47 crc kubenswrapper[4678]: E1206 11:06:47.477253 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:07:01 crc kubenswrapper[4678]: I1206 11:07:01.476632 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:07:01 crc kubenswrapper[4678]: E1206 11:07:01.477731 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:07:16 crc kubenswrapper[4678]: I1206 11:07:16.479846 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:07:16 crc kubenswrapper[4678]: E1206 11:07:16.480938 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:07:23 crc kubenswrapper[4678]: I1206 11:07:23.033784 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-6btts"] Dec 06 11:07:23 crc kubenswrapper[4678]: I1206 11:07:23.040901 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-6btts"] Dec 06 11:07:23 crc kubenswrapper[4678]: I1206 11:07:23.488810 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a54c6d51-0b8e-491d-b5fb-8d08a260232a" path="/var/lib/kubelet/pods/a54c6d51-0b8e-491d-b5fb-8d08a260232a/volumes" Dec 06 11:07:24 crc kubenswrapper[4678]: I1206 11:07:24.026663 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-39d4-account-create-update-wpqkc"] Dec 06 11:07:24 crc kubenswrapper[4678]: I1206 11:07:24.040067 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-39d4-account-create-update-wpqkc"] Dec 06 11:07:25 crc kubenswrapper[4678]: I1206 11:07:25.496000 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c3dd118-3077-4c96-b13b-edb34a16d944" path="/var/lib/kubelet/pods/1c3dd118-3077-4c96-b13b-edb34a16d944/volumes" Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.027157 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-4zkvd"] Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.037422 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-455a-account-create-update-kg2j5"] Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.051961 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7391-account-create-update-74ks6"] Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.064890 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-8kxvz"] Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.075084 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-455a-account-create-update-kg2j5"] Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.083819 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-4zkvd"] Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.091277 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7391-account-create-update-74ks6"] Dec 06 11:07:26 crc kubenswrapper[4678]: I1206 11:07:26.098153 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-8kxvz"] Dec 06 11:07:27 crc kubenswrapper[4678]: I1206 11:07:27.477127 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:07:27 crc kubenswrapper[4678]: E1206 11:07:27.477715 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:07:27 crc kubenswrapper[4678]: I1206 11:07:27.489147 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="081454ce-456d-4cd3-bec3-b7a06034ab39" path="/var/lib/kubelet/pods/081454ce-456d-4cd3-bec3-b7a06034ab39/volumes" Dec 06 11:07:27 crc kubenswrapper[4678]: I1206 11:07:27.489746 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="582eb659-cea1-44f9-9d87-a2d965a9bcfc" path="/var/lib/kubelet/pods/582eb659-cea1-44f9-9d87-a2d965a9bcfc/volumes" Dec 06 11:07:27 crc kubenswrapper[4678]: I1206 11:07:27.490277 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee" path="/var/lib/kubelet/pods/cfc9e1e3-0821-4aaf-bc65-ad425b6fe9ee/volumes" Dec 06 11:07:27 crc kubenswrapper[4678]: I1206 11:07:27.490844 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d549492c-fcf7-402b-95e5-c0bee52f86eb" path="/var/lib/kubelet/pods/d549492c-fcf7-402b-95e5-c0bee52f86eb/volumes" Dec 06 11:07:40 crc kubenswrapper[4678]: I1206 11:07:40.476372 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:07:40 crc kubenswrapper[4678]: E1206 11:07:40.477312 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:07:46 crc kubenswrapper[4678]: I1206 11:07:46.893700 4678 scope.go:117] "RemoveContainer" containerID="913a209132c2adb82e0be8f3c4fd4db2fefc66393db79043d32d41e49e286137" Dec 06 11:07:46 crc kubenswrapper[4678]: I1206 11:07:46.924662 4678 scope.go:117] "RemoveContainer" containerID="7024ffc23101ee98af7e162435be201e29c64be1737c12844284f2d9ed2950f6" Dec 06 11:07:46 crc kubenswrapper[4678]: I1206 11:07:46.973643 4678 scope.go:117] "RemoveContainer" containerID="2fd015cd365d86d0f36cd85f92c016d03e9866baea84532882fda84116b71284" Dec 06 11:07:47 crc kubenswrapper[4678]: I1206 11:07:47.014082 4678 scope.go:117] "RemoveContainer" containerID="f226fac5ff92d5521c8126125c7ea7b1f138f5f0939b61fbd9643a0c40b5741f" Dec 06 11:07:47 crc kubenswrapper[4678]: I1206 11:07:47.050998 4678 scope.go:117] "RemoveContainer" containerID="7c02b4470a48fb0bbae65eb0738605e21260798f2505835287da1c0d656eca4e" Dec 06 11:07:47 crc kubenswrapper[4678]: I1206 11:07:47.096338 4678 scope.go:117] "RemoveContainer" containerID="bdfaabcfa35fb516f840b0a76fcd3749e0e5ec8ddfd8f9a6ef6825000f7bdc3b" Dec 06 11:07:52 crc kubenswrapper[4678]: I1206 11:07:52.476691 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:07:52 crc kubenswrapper[4678]: E1206 11:07:52.477595 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:08:00 crc kubenswrapper[4678]: I1206 11:08:00.549622 4678 generic.go:334] "Generic (PLEG): container finished" podID="5fa31310-7e0f-431a-ae13-287fd25a926c" containerID="9bb2618347ba4c503d1285af113fc9d7d2fa61fd7c0fac4678a1e6d2fbce1d4a" exitCode=0 Dec 06 11:08:00 crc kubenswrapper[4678]: I1206 11:08:00.549705 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" event={"ID":"5fa31310-7e0f-431a-ae13-287fd25a926c","Type":"ContainerDied","Data":"9bb2618347ba4c503d1285af113fc9d7d2fa61fd7c0fac4678a1e6d2fbce1d4a"} Dec 06 11:08:01 crc kubenswrapper[4678]: I1206 11:08:01.973998 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.051000 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-inventory\") pod \"5fa31310-7e0f-431a-ae13-287fd25a926c\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.051075 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-ssh-key\") pod \"5fa31310-7e0f-431a-ae13-287fd25a926c\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.051132 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjdgn\" (UniqueName: \"kubernetes.io/projected/5fa31310-7e0f-431a-ae13-287fd25a926c-kube-api-access-mjdgn\") pod \"5fa31310-7e0f-431a-ae13-287fd25a926c\" (UID: \"5fa31310-7e0f-431a-ae13-287fd25a926c\") " Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.064246 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa31310-7e0f-431a-ae13-287fd25a926c-kube-api-access-mjdgn" (OuterVolumeSpecName: "kube-api-access-mjdgn") pod "5fa31310-7e0f-431a-ae13-287fd25a926c" (UID: "5fa31310-7e0f-431a-ae13-287fd25a926c"). InnerVolumeSpecName "kube-api-access-mjdgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.083219 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5fa31310-7e0f-431a-ae13-287fd25a926c" (UID: "5fa31310-7e0f-431a-ae13-287fd25a926c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.083369 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-inventory" (OuterVolumeSpecName: "inventory") pod "5fa31310-7e0f-431a-ae13-287fd25a926c" (UID: "5fa31310-7e0f-431a-ae13-287fd25a926c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.153123 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.153154 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5fa31310-7e0f-431a-ae13-287fd25a926c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.153167 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjdgn\" (UniqueName: \"kubernetes.io/projected/5fa31310-7e0f-431a-ae13-287fd25a926c-kube-api-access-mjdgn\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.571641 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" event={"ID":"5fa31310-7e0f-431a-ae13-287fd25a926c","Type":"ContainerDied","Data":"2ee4c19db692dd8f7f2cfffa8d6b640a9971610ed2f0770d65e521dc2e9a5d2a"} Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.571970 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ee4c19db692dd8f7f2cfffa8d6b640a9971610ed2f0770d65e521dc2e9a5d2a" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.571680 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.670703 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs"] Dec 06 11:08:02 crc kubenswrapper[4678]: E1206 11:08:02.671954 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa31310-7e0f-431a-ae13-287fd25a926c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.671979 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa31310-7e0f-431a-ae13-287fd25a926c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.672211 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa31310-7e0f-431a-ae13-287fd25a926c" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.672964 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.678047 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.678266 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.678279 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.678451 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.684542 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs"] Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.764549 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.764614 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.764697 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vp27q\" (UniqueName: \"kubernetes.io/projected/89741e1b-d65e-4e23-982d-716c2e498c25-kube-api-access-vp27q\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.866673 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.866769 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vp27q\" (UniqueName: \"kubernetes.io/projected/89741e1b-d65e-4e23-982d-716c2e498c25-kube-api-access-vp27q\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.866939 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.873233 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.876045 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:02 crc kubenswrapper[4678]: I1206 11:08:02.884264 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vp27q\" (UniqueName: \"kubernetes.io/projected/89741e1b-d65e-4e23-982d-716c2e498c25-kube-api-access-vp27q\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:03 crc kubenswrapper[4678]: I1206 11:08:03.002556 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:03 crc kubenswrapper[4678]: I1206 11:08:03.475573 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:08:03 crc kubenswrapper[4678]: I1206 11:08:03.590973 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs"] Dec 06 11:08:04 crc kubenswrapper[4678]: I1206 11:08:04.590204 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"8b5b29cf01b4f5ab8126b1632136a603737bbd297541220236e9787fb95d5818"} Dec 06 11:08:04 crc kubenswrapper[4678]: I1206 11:08:04.593644 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" event={"ID":"89741e1b-d65e-4e23-982d-716c2e498c25","Type":"ContainerStarted","Data":"d53196856c077c166aa9275987dc5c6a7bdcc7cd95d78e3a652f7fdf3ff60895"} Dec 06 11:08:04 crc kubenswrapper[4678]: I1206 11:08:04.593681 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" event={"ID":"89741e1b-d65e-4e23-982d-716c2e498c25","Type":"ContainerStarted","Data":"47564f24d9f4f58e66543f7442429bca343966b81921d220332319b4fac9896a"} Dec 06 11:08:04 crc kubenswrapper[4678]: I1206 11:08:04.640162 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" podStartSLOduration=2.180601414 podStartE2EDuration="2.640142028s" podCreationTimestamp="2025-12-06 11:08:02 +0000 UTC" firstStartedPulling="2025-12-06 11:08:03.603680896 +0000 UTC m=+1888.447112345" lastFinishedPulling="2025-12-06 11:08:04.06322151 +0000 UTC m=+1888.906652959" observedRunningTime="2025-12-06 11:08:04.622311848 +0000 UTC m=+1889.465743287" watchObservedRunningTime="2025-12-06 11:08:04.640142028 +0000 UTC m=+1889.483573457" Dec 06 11:08:09 crc kubenswrapper[4678]: I1206 11:08:09.643631 4678 generic.go:334] "Generic (PLEG): container finished" podID="89741e1b-d65e-4e23-982d-716c2e498c25" containerID="d53196856c077c166aa9275987dc5c6a7bdcc7cd95d78e3a652f7fdf3ff60895" exitCode=0 Dec 06 11:08:09 crc kubenswrapper[4678]: I1206 11:08:09.643710 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" event={"ID":"89741e1b-d65e-4e23-982d-716c2e498c25","Type":"ContainerDied","Data":"d53196856c077c166aa9275987dc5c6a7bdcc7cd95d78e3a652f7fdf3ff60895"} Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.042580 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.126835 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-ssh-key\") pod \"89741e1b-d65e-4e23-982d-716c2e498c25\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.127089 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-inventory\") pod \"89741e1b-d65e-4e23-982d-716c2e498c25\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.127162 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vp27q\" (UniqueName: \"kubernetes.io/projected/89741e1b-d65e-4e23-982d-716c2e498c25-kube-api-access-vp27q\") pod \"89741e1b-d65e-4e23-982d-716c2e498c25\" (UID: \"89741e1b-d65e-4e23-982d-716c2e498c25\") " Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.132794 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89741e1b-d65e-4e23-982d-716c2e498c25-kube-api-access-vp27q" (OuterVolumeSpecName: "kube-api-access-vp27q") pod "89741e1b-d65e-4e23-982d-716c2e498c25" (UID: "89741e1b-d65e-4e23-982d-716c2e498c25"). InnerVolumeSpecName "kube-api-access-vp27q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.158225 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-inventory" (OuterVolumeSpecName: "inventory") pod "89741e1b-d65e-4e23-982d-716c2e498c25" (UID: "89741e1b-d65e-4e23-982d-716c2e498c25"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.169134 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "89741e1b-d65e-4e23-982d-716c2e498c25" (UID: "89741e1b-d65e-4e23-982d-716c2e498c25"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.229816 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.230041 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vp27q\" (UniqueName: \"kubernetes.io/projected/89741e1b-d65e-4e23-982d-716c2e498c25-kube-api-access-vp27q\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.230128 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/89741e1b-d65e-4e23-982d-716c2e498c25-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.667931 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" event={"ID":"89741e1b-d65e-4e23-982d-716c2e498c25","Type":"ContainerDied","Data":"47564f24d9f4f58e66543f7442429bca343966b81921d220332319b4fac9896a"} Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.667971 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47564f24d9f4f58e66543f7442429bca343966b81921d220332319b4fac9896a" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.668031 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.751818 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m"] Dec 06 11:08:11 crc kubenswrapper[4678]: E1206 11:08:11.752481 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89741e1b-d65e-4e23-982d-716c2e498c25" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.752522 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="89741e1b-d65e-4e23-982d-716c2e498c25" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.752695 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="89741e1b-d65e-4e23-982d-716c2e498c25" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.753306 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.756399 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.756872 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.757996 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.764290 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m"] Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.798564 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.839435 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.840036 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.840982 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvlh7\" (UniqueName: \"kubernetes.io/projected/8e6b45b8-d8db-457e-9a08-04ce67b197f6-kube-api-access-jvlh7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.943090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.943421 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.943599 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvlh7\" (UniqueName: \"kubernetes.io/projected/8e6b45b8-d8db-457e-9a08-04ce67b197f6-kube-api-access-jvlh7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.948885 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.952215 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:11 crc kubenswrapper[4678]: I1206 11:08:11.962082 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvlh7\" (UniqueName: \"kubernetes.io/projected/8e6b45b8-d8db-457e-9a08-04ce67b197f6-kube-api-access-jvlh7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4l22m\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:12 crc kubenswrapper[4678]: I1206 11:08:12.103576 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:12 crc kubenswrapper[4678]: I1206 11:08:12.599170 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m"] Dec 06 11:08:12 crc kubenswrapper[4678]: I1206 11:08:12.676931 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" event={"ID":"8e6b45b8-d8db-457e-9a08-04ce67b197f6","Type":"ContainerStarted","Data":"f441005106229a7c46d16227cf40219fb80044d24cec5ef9986d32f076f842db"} Dec 06 11:08:13 crc kubenswrapper[4678]: I1206 11:08:13.688617 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" event={"ID":"8e6b45b8-d8db-457e-9a08-04ce67b197f6","Type":"ContainerStarted","Data":"c70da411cc31ffb62edceb43e300bb34d5c4a49a890d5ae82fa6a98098818e56"} Dec 06 11:08:13 crc kubenswrapper[4678]: I1206 11:08:13.721579 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" podStartSLOduration=2.308813178 podStartE2EDuration="2.721545067s" podCreationTimestamp="2025-12-06 11:08:11 +0000 UTC" firstStartedPulling="2025-12-06 11:08:12.602954901 +0000 UTC m=+1897.446386340" lastFinishedPulling="2025-12-06 11:08:13.01568679 +0000 UTC m=+1897.859118229" observedRunningTime="2025-12-06 11:08:13.719141041 +0000 UTC m=+1898.562572520" watchObservedRunningTime="2025-12-06 11:08:13.721545067 +0000 UTC m=+1898.564976546" Dec 06 11:08:14 crc kubenswrapper[4678]: I1206 11:08:14.048413 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdm9z"] Dec 06 11:08:14 crc kubenswrapper[4678]: I1206 11:08:14.058919 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xdm9z"] Dec 06 11:08:15 crc kubenswrapper[4678]: I1206 11:08:15.491003 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d1feb10-f542-4553-89fc-373d1940ac22" path="/var/lib/kubelet/pods/4d1feb10-f542-4553-89fc-373d1940ac22/volumes" Dec 06 11:08:36 crc kubenswrapper[4678]: I1206 11:08:36.038102 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-thzqq"] Dec 06 11:08:36 crc kubenswrapper[4678]: I1206 11:08:36.048255 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-thzqq"] Dec 06 11:08:37 crc kubenswrapper[4678]: I1206 11:08:37.487798 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dea23b4d-e666-4ba6-954d-c07956044128" path="/var/lib/kubelet/pods/dea23b4d-e666-4ba6-954d-c07956044128/volumes" Dec 06 11:08:40 crc kubenswrapper[4678]: I1206 11:08:40.034220 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-df7h2"] Dec 06 11:08:40 crc kubenswrapper[4678]: I1206 11:08:40.042064 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-df7h2"] Dec 06 11:08:41 crc kubenswrapper[4678]: I1206 11:08:41.490882 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02942adf-5e19-482e-a76a-16d40ec52416" path="/var/lib/kubelet/pods/02942adf-5e19-482e-a76a-16d40ec52416/volumes" Dec 06 11:08:47 crc kubenswrapper[4678]: I1206 11:08:47.249240 4678 scope.go:117] "RemoveContainer" containerID="74c7cee84aa2b1df3deb15ba7c039b13fa352a76d0712b6dda21236564b9dced" Dec 06 11:08:47 crc kubenswrapper[4678]: I1206 11:08:47.282205 4678 scope.go:117] "RemoveContainer" containerID="486a5433f3fcd5aa9866fb1bba4921184592cda13ee965cc3dbd945e0290f2a2" Dec 06 11:08:47 crc kubenswrapper[4678]: I1206 11:08:47.330259 4678 scope.go:117] "RemoveContainer" containerID="25ffb982290c4206502d8c86b1733a1e6c83d918b7c462e18e09ff55e90541f3" Dec 06 11:08:47 crc kubenswrapper[4678]: I1206 11:08:47.399544 4678 scope.go:117] "RemoveContainer" containerID="f5dedb8605288de355a8c5985e6316ff3a4a5e66a0ad788424e803cd26e0b01d" Dec 06 11:08:47 crc kubenswrapper[4678]: I1206 11:08:47.428154 4678 scope.go:117] "RemoveContainer" containerID="0b127b3a998bc68c192239209fdbf6fee3765fd1f2d5bb7a1fd29011bbf759b7" Dec 06 11:08:47 crc kubenswrapper[4678]: I1206 11:08:47.478704 4678 scope.go:117] "RemoveContainer" containerID="cd9f48ba61fb865dceb1d3b0e300a2f28ad2bb7a553679d3b216a45aaee01e08" Dec 06 11:08:58 crc kubenswrapper[4678]: I1206 11:08:58.104998 4678 generic.go:334] "Generic (PLEG): container finished" podID="8e6b45b8-d8db-457e-9a08-04ce67b197f6" containerID="c70da411cc31ffb62edceb43e300bb34d5c4a49a890d5ae82fa6a98098818e56" exitCode=0 Dec 06 11:08:58 crc kubenswrapper[4678]: I1206 11:08:58.105196 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" event={"ID":"8e6b45b8-d8db-457e-9a08-04ce67b197f6","Type":"ContainerDied","Data":"c70da411cc31ffb62edceb43e300bb34d5c4a49a890d5ae82fa6a98098818e56"} Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.545596 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.593568 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvlh7\" (UniqueName: \"kubernetes.io/projected/8e6b45b8-d8db-457e-9a08-04ce67b197f6-kube-api-access-jvlh7\") pod \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.593670 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-ssh-key\") pod \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.593813 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-inventory\") pod \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\" (UID: \"8e6b45b8-d8db-457e-9a08-04ce67b197f6\") " Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.604206 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6b45b8-d8db-457e-9a08-04ce67b197f6-kube-api-access-jvlh7" (OuterVolumeSpecName: "kube-api-access-jvlh7") pod "8e6b45b8-d8db-457e-9a08-04ce67b197f6" (UID: "8e6b45b8-d8db-457e-9a08-04ce67b197f6"). InnerVolumeSpecName "kube-api-access-jvlh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.651094 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-inventory" (OuterVolumeSpecName: "inventory") pod "8e6b45b8-d8db-457e-9a08-04ce67b197f6" (UID: "8e6b45b8-d8db-457e-9a08-04ce67b197f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.652471 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8e6b45b8-d8db-457e-9a08-04ce67b197f6" (UID: "8e6b45b8-d8db-457e-9a08-04ce67b197f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.696040 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvlh7\" (UniqueName: \"kubernetes.io/projected/8e6b45b8-d8db-457e-9a08-04ce67b197f6-kube-api-access-jvlh7\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.696076 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:08:59 crc kubenswrapper[4678]: I1206 11:08:59.696086 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8e6b45b8-d8db-457e-9a08-04ce67b197f6-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.132506 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" event={"ID":"8e6b45b8-d8db-457e-9a08-04ce67b197f6","Type":"ContainerDied","Data":"f441005106229a7c46d16227cf40219fb80044d24cec5ef9986d32f076f842db"} Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.132799 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f441005106229a7c46d16227cf40219fb80044d24cec5ef9986d32f076f842db" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.132558 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4l22m" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.227214 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8"] Dec 06 11:09:00 crc kubenswrapper[4678]: E1206 11:09:00.227677 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6b45b8-d8db-457e-9a08-04ce67b197f6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.227696 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6b45b8-d8db-457e-9a08-04ce67b197f6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.227920 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6b45b8-d8db-457e-9a08-04ce67b197f6" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.228807 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.234033 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.234144 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.234206 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.234345 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.256870 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8"] Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.408847 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.408953 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7wnl\" (UniqueName: \"kubernetes.io/projected/f104b9bb-12fa-4f5c-86ed-2db61119cb55-kube-api-access-t7wnl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.408986 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.511941 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.512069 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7wnl\" (UniqueName: \"kubernetes.io/projected/f104b9bb-12fa-4f5c-86ed-2db61119cb55-kube-api-access-t7wnl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.512118 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.517187 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.525559 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.529973 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7wnl\" (UniqueName: \"kubernetes.io/projected/f104b9bb-12fa-4f5c-86ed-2db61119cb55-kube-api-access-t7wnl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:00 crc kubenswrapper[4678]: I1206 11:09:00.561429 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:09:01 crc kubenswrapper[4678]: I1206 11:09:01.311673 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8"] Dec 06 11:09:02 crc kubenswrapper[4678]: I1206 11:09:02.152133 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" event={"ID":"f104b9bb-12fa-4f5c-86ed-2db61119cb55","Type":"ContainerStarted","Data":"3bde5b1cc989463457826a5048017ab759fcb3ebb5a9fa92ea4b91119a61398d"} Dec 06 11:09:02 crc kubenswrapper[4678]: I1206 11:09:02.152664 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" event={"ID":"f104b9bb-12fa-4f5c-86ed-2db61119cb55","Type":"ContainerStarted","Data":"e3c3835e2c78eb9477f5b93da4ec42a4d5f4a249fe4cf1f4cb4139f5f27ecf67"} Dec 06 11:09:02 crc kubenswrapper[4678]: I1206 11:09:02.229964 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" podStartSLOduration=1.779545448 podStartE2EDuration="2.229943674s" podCreationTimestamp="2025-12-06 11:09:00 +0000 UTC" firstStartedPulling="2025-12-06 11:09:01.31764339 +0000 UTC m=+1946.161074829" lastFinishedPulling="2025-12-06 11:09:01.768041616 +0000 UTC m=+1946.611473055" observedRunningTime="2025-12-06 11:09:02.2266379 +0000 UTC m=+1947.070069339" watchObservedRunningTime="2025-12-06 11:09:02.229943674 +0000 UTC m=+1947.073375123" Dec 06 11:09:18 crc kubenswrapper[4678]: I1206 11:09:18.050312 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9wvrw"] Dec 06 11:09:18 crc kubenswrapper[4678]: I1206 11:09:18.060548 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9wvrw"] Dec 06 11:09:19 crc kubenswrapper[4678]: I1206 11:09:19.485591 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a26e1f26-4f6d-49c4-a66a-1ca2fc519d46" path="/var/lib/kubelet/pods/a26e1f26-4f6d-49c4-a66a-1ca2fc519d46/volumes" Dec 06 11:09:47 crc kubenswrapper[4678]: I1206 11:09:47.583742 4678 scope.go:117] "RemoveContainer" containerID="4078ef8e9e930b90e52963fd33488744b48575ad03bc0edee8ee104b5cb41795" Dec 06 11:10:04 crc kubenswrapper[4678]: I1206 11:10:04.784076 4678 generic.go:334] "Generic (PLEG): container finished" podID="f104b9bb-12fa-4f5c-86ed-2db61119cb55" containerID="3bde5b1cc989463457826a5048017ab759fcb3ebb5a9fa92ea4b91119a61398d" exitCode=0 Dec 06 11:10:04 crc kubenswrapper[4678]: I1206 11:10:04.784289 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" event={"ID":"f104b9bb-12fa-4f5c-86ed-2db61119cb55","Type":"ContainerDied","Data":"3bde5b1cc989463457826a5048017ab759fcb3ebb5a9fa92ea4b91119a61398d"} Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.146983 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.317610 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-inventory\") pod \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.317671 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7wnl\" (UniqueName: \"kubernetes.io/projected/f104b9bb-12fa-4f5c-86ed-2db61119cb55-kube-api-access-t7wnl\") pod \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.317757 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-ssh-key\") pod \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\" (UID: \"f104b9bb-12fa-4f5c-86ed-2db61119cb55\") " Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.334110 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f104b9bb-12fa-4f5c-86ed-2db61119cb55-kube-api-access-t7wnl" (OuterVolumeSpecName: "kube-api-access-t7wnl") pod "f104b9bb-12fa-4f5c-86ed-2db61119cb55" (UID: "f104b9bb-12fa-4f5c-86ed-2db61119cb55"). InnerVolumeSpecName "kube-api-access-t7wnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.344544 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f104b9bb-12fa-4f5c-86ed-2db61119cb55" (UID: "f104b9bb-12fa-4f5c-86ed-2db61119cb55"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.367302 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-inventory" (OuterVolumeSpecName: "inventory") pod "f104b9bb-12fa-4f5c-86ed-2db61119cb55" (UID: "f104b9bb-12fa-4f5c-86ed-2db61119cb55"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.420412 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.420467 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7wnl\" (UniqueName: \"kubernetes.io/projected/f104b9bb-12fa-4f5c-86ed-2db61119cb55-kube-api-access-t7wnl\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.420497 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f104b9bb-12fa-4f5c-86ed-2db61119cb55-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.805670 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" event={"ID":"f104b9bb-12fa-4f5c-86ed-2db61119cb55","Type":"ContainerDied","Data":"e3c3835e2c78eb9477f5b93da4ec42a4d5f4a249fe4cf1f4cb4139f5f27ecf67"} Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.805950 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3c3835e2c78eb9477f5b93da4ec42a4d5f4a249fe4cf1f4cb4139f5f27ecf67" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.805756 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.916124 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9pvlf"] Dec 06 11:10:06 crc kubenswrapper[4678]: E1206 11:10:06.916684 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f104b9bb-12fa-4f5c-86ed-2db61119cb55" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.916707 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="f104b9bb-12fa-4f5c-86ed-2db61119cb55" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.916892 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="f104b9bb-12fa-4f5c-86ed-2db61119cb55" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.917626 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.920902 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.921057 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.921215 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.932034 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9pvlf"] Dec 06 11:10:06 crc kubenswrapper[4678]: I1206 11:10:06.932157 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.031972 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gngzp\" (UniqueName: \"kubernetes.io/projected/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-kube-api-access-gngzp\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.032165 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.032484 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.135245 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gngzp\" (UniqueName: \"kubernetes.io/projected/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-kube-api-access-gngzp\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.135368 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.135593 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.142286 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.142553 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.162131 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gngzp\" (UniqueName: \"kubernetes.io/projected/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-kube-api-access-gngzp\") pod \"ssh-known-hosts-edpm-deployment-9pvlf\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.282383 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.852116 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:10:07 crc kubenswrapper[4678]: I1206 11:10:07.855462 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-9pvlf"] Dec 06 11:10:08 crc kubenswrapper[4678]: I1206 11:10:08.839882 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" event={"ID":"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed","Type":"ContainerStarted","Data":"5bc54273ebff04a688acf0b88b9f16bf7e7a2a150e0b13b36a6ed387e0b13250"} Dec 06 11:10:08 crc kubenswrapper[4678]: I1206 11:10:08.840951 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" event={"ID":"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed","Type":"ContainerStarted","Data":"a6cf256e63fe7c2f7542b47011bd82508befbe9a989c8bdd5b13ff0f210a805d"} Dec 06 11:10:08 crc kubenswrapper[4678]: I1206 11:10:08.863614 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" podStartSLOduration=2.296603061 podStartE2EDuration="2.863594947s" podCreationTimestamp="2025-12-06 11:10:06 +0000 UTC" firstStartedPulling="2025-12-06 11:10:07.851861094 +0000 UTC m=+2012.695292533" lastFinishedPulling="2025-12-06 11:10:08.41885297 +0000 UTC m=+2013.262284419" observedRunningTime="2025-12-06 11:10:08.858816987 +0000 UTC m=+2013.702248426" watchObservedRunningTime="2025-12-06 11:10:08.863594947 +0000 UTC m=+2013.707026386" Dec 06 11:10:16 crc kubenswrapper[4678]: I1206 11:10:16.911467 4678 generic.go:334] "Generic (PLEG): container finished" podID="06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed" containerID="5bc54273ebff04a688acf0b88b9f16bf7e7a2a150e0b13b36a6ed387e0b13250" exitCode=0 Dec 06 11:10:16 crc kubenswrapper[4678]: I1206 11:10:16.912022 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" event={"ID":"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed","Type":"ContainerDied","Data":"5bc54273ebff04a688acf0b88b9f16bf7e7a2a150e0b13b36a6ed387e0b13250"} Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.322449 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.478146 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-inventory-0\") pod \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.478252 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gngzp\" (UniqueName: \"kubernetes.io/projected/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-kube-api-access-gngzp\") pod \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.478294 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-ssh-key-openstack-edpm-ipam\") pod \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\" (UID: \"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed\") " Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.483748 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-kube-api-access-gngzp" (OuterVolumeSpecName: "kube-api-access-gngzp") pod "06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed" (UID: "06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed"). InnerVolumeSpecName "kube-api-access-gngzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.505278 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed" (UID: "06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.514451 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed" (UID: "06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.581132 4678 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.581194 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gngzp\" (UniqueName: \"kubernetes.io/projected/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-kube-api-access-gngzp\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.581210 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.933837 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" event={"ID":"06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed","Type":"ContainerDied","Data":"a6cf256e63fe7c2f7542b47011bd82508befbe9a989c8bdd5b13ff0f210a805d"} Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.933881 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6cf256e63fe7c2f7542b47011bd82508befbe9a989c8bdd5b13ff0f210a805d" Dec 06 11:10:18 crc kubenswrapper[4678]: I1206 11:10:18.934244 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-9pvlf" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.040295 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq"] Dec 06 11:10:19 crc kubenswrapper[4678]: E1206 11:10:19.040744 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed" containerName="ssh-known-hosts-edpm-deployment" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.040758 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed" containerName="ssh-known-hosts-edpm-deployment" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.040977 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed" containerName="ssh-known-hosts-edpm-deployment" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.041653 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.046163 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.046347 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.046683 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.046712 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.055823 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq"] Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.197899 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.198054 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgg4v\" (UniqueName: \"kubernetes.io/projected/fdc4c25f-dc36-4a52-a26d-9dd390c23061-kube-api-access-rgg4v\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.198110 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.300180 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.300664 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgg4v\" (UniqueName: \"kubernetes.io/projected/fdc4c25f-dc36-4a52-a26d-9dd390c23061-kube-api-access-rgg4v\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.300844 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.304975 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.304988 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.321915 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgg4v\" (UniqueName: \"kubernetes.io/projected/fdc4c25f-dc36-4a52-a26d-9dd390c23061-kube-api-access-rgg4v\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-7kqwq\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.363403 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:19 crc kubenswrapper[4678]: I1206 11:10:19.991654 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq"] Dec 06 11:10:20 crc kubenswrapper[4678]: I1206 11:10:20.952697 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" event={"ID":"fdc4c25f-dc36-4a52-a26d-9dd390c23061","Type":"ContainerStarted","Data":"15ef6bf002a839c798035cc5338938ea49f2fc95475950102779ccf4003bd0ee"} Dec 06 11:10:20 crc kubenswrapper[4678]: I1206 11:10:20.954109 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" event={"ID":"fdc4c25f-dc36-4a52-a26d-9dd390c23061","Type":"ContainerStarted","Data":"237dfd0c56eca7b3c735772bfc1471c28fb25b91ab7cd7915c89f3d465bc9614"} Dec 06 11:10:20 crc kubenswrapper[4678]: I1206 11:10:20.971804 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" podStartSLOduration=1.539338662 podStartE2EDuration="1.971785312s" podCreationTimestamp="2025-12-06 11:10:19 +0000 UTC" firstStartedPulling="2025-12-06 11:10:19.995302008 +0000 UTC m=+2024.838733447" lastFinishedPulling="2025-12-06 11:10:20.427748658 +0000 UTC m=+2025.271180097" observedRunningTime="2025-12-06 11:10:20.969073136 +0000 UTC m=+2025.812504595" watchObservedRunningTime="2025-12-06 11:10:20.971785312 +0000 UTC m=+2025.815216761" Dec 06 11:10:29 crc kubenswrapper[4678]: I1206 11:10:29.505627 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:10:29 crc kubenswrapper[4678]: I1206 11:10:29.505905 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:10:30 crc kubenswrapper[4678]: I1206 11:10:30.046972 4678 generic.go:334] "Generic (PLEG): container finished" podID="fdc4c25f-dc36-4a52-a26d-9dd390c23061" containerID="15ef6bf002a839c798035cc5338938ea49f2fc95475950102779ccf4003bd0ee" exitCode=0 Dec 06 11:10:30 crc kubenswrapper[4678]: I1206 11:10:30.047055 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" event={"ID":"fdc4c25f-dc36-4a52-a26d-9dd390c23061","Type":"ContainerDied","Data":"15ef6bf002a839c798035cc5338938ea49f2fc95475950102779ccf4003bd0ee"} Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.501906 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.655892 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-ssh-key\") pod \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.656172 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-inventory\") pod \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.656196 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgg4v\" (UniqueName: \"kubernetes.io/projected/fdc4c25f-dc36-4a52-a26d-9dd390c23061-kube-api-access-rgg4v\") pod \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\" (UID: \"fdc4c25f-dc36-4a52-a26d-9dd390c23061\") " Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.662823 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdc4c25f-dc36-4a52-a26d-9dd390c23061-kube-api-access-rgg4v" (OuterVolumeSpecName: "kube-api-access-rgg4v") pod "fdc4c25f-dc36-4a52-a26d-9dd390c23061" (UID: "fdc4c25f-dc36-4a52-a26d-9dd390c23061"). InnerVolumeSpecName "kube-api-access-rgg4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.683582 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fdc4c25f-dc36-4a52-a26d-9dd390c23061" (UID: "fdc4c25f-dc36-4a52-a26d-9dd390c23061"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.684716 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-inventory" (OuterVolumeSpecName: "inventory") pod "fdc4c25f-dc36-4a52-a26d-9dd390c23061" (UID: "fdc4c25f-dc36-4a52-a26d-9dd390c23061"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.758011 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.758053 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgg4v\" (UniqueName: \"kubernetes.io/projected/fdc4c25f-dc36-4a52-a26d-9dd390c23061-kube-api-access-rgg4v\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:31 crc kubenswrapper[4678]: I1206 11:10:31.758068 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fdc4c25f-dc36-4a52-a26d-9dd390c23061-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.083173 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" event={"ID":"fdc4c25f-dc36-4a52-a26d-9dd390c23061","Type":"ContainerDied","Data":"237dfd0c56eca7b3c735772bfc1471c28fb25b91ab7cd7915c89f3d465bc9614"} Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.083209 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="237dfd0c56eca7b3c735772bfc1471c28fb25b91ab7cd7915c89f3d465bc9614" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.083245 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-7kqwq" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.157560 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz"] Dec 06 11:10:32 crc kubenswrapper[4678]: E1206 11:10:32.157934 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc4c25f-dc36-4a52-a26d-9dd390c23061" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.157950 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc4c25f-dc36-4a52-a26d-9dd390c23061" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.158183 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdc4c25f-dc36-4a52-a26d-9dd390c23061" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.158809 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.166485 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.166984 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.167552 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.167726 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.171814 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.171955 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sff2q\" (UniqueName: \"kubernetes.io/projected/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-kube-api-access-sff2q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.172195 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.179351 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz"] Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.274100 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.274175 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sff2q\" (UniqueName: \"kubernetes.io/projected/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-kube-api-access-sff2q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.274939 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.277654 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.278634 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.292678 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sff2q\" (UniqueName: \"kubernetes.io/projected/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-kube-api-access-sff2q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:32 crc kubenswrapper[4678]: I1206 11:10:32.527216 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:33 crc kubenswrapper[4678]: I1206 11:10:33.048451 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz"] Dec 06 11:10:33 crc kubenswrapper[4678]: I1206 11:10:33.091608 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" event={"ID":"d6e6e2e2-4745-46f6-a3fa-94f803ac0677","Type":"ContainerStarted","Data":"0590be8f2dd0918d1c795b32da9ded9231a5dbe787a69a5cc32499b22347b84d"} Dec 06 11:10:34 crc kubenswrapper[4678]: I1206 11:10:34.107690 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" event={"ID":"d6e6e2e2-4745-46f6-a3fa-94f803ac0677","Type":"ContainerStarted","Data":"7741eb96e254051d1287af9658466c7a5a152431e5a8fba0e1a2e351d70641ba"} Dec 06 11:10:34 crc kubenswrapper[4678]: I1206 11:10:34.155357 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" podStartSLOduration=1.755067734 podStartE2EDuration="2.155331312s" podCreationTimestamp="2025-12-06 11:10:32 +0000 UTC" firstStartedPulling="2025-12-06 11:10:33.048875077 +0000 UTC m=+2037.892306516" lastFinishedPulling="2025-12-06 11:10:33.449138645 +0000 UTC m=+2038.292570094" observedRunningTime="2025-12-06 11:10:34.149929522 +0000 UTC m=+2038.993360981" watchObservedRunningTime="2025-12-06 11:10:34.155331312 +0000 UTC m=+2038.998762751" Dec 06 11:10:44 crc kubenswrapper[4678]: I1206 11:10:44.194143 4678 generic.go:334] "Generic (PLEG): container finished" podID="d6e6e2e2-4745-46f6-a3fa-94f803ac0677" containerID="7741eb96e254051d1287af9658466c7a5a152431e5a8fba0e1a2e351d70641ba" exitCode=0 Dec 06 11:10:44 crc kubenswrapper[4678]: I1206 11:10:44.194183 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" event={"ID":"d6e6e2e2-4745-46f6-a3fa-94f803ac0677","Type":"ContainerDied","Data":"7741eb96e254051d1287af9658466c7a5a152431e5a8fba0e1a2e351d70641ba"} Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.699563 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.748573 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-inventory\") pod \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.748699 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sff2q\" (UniqueName: \"kubernetes.io/projected/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-kube-api-access-sff2q\") pod \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.748733 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-ssh-key\") pod \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\" (UID: \"d6e6e2e2-4745-46f6-a3fa-94f803ac0677\") " Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.762854 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-kube-api-access-sff2q" (OuterVolumeSpecName: "kube-api-access-sff2q") pod "d6e6e2e2-4745-46f6-a3fa-94f803ac0677" (UID: "d6e6e2e2-4745-46f6-a3fa-94f803ac0677"). InnerVolumeSpecName "kube-api-access-sff2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.779895 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d6e6e2e2-4745-46f6-a3fa-94f803ac0677" (UID: "d6e6e2e2-4745-46f6-a3fa-94f803ac0677"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.780713 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-inventory" (OuterVolumeSpecName: "inventory") pod "d6e6e2e2-4745-46f6-a3fa-94f803ac0677" (UID: "d6e6e2e2-4745-46f6-a3fa-94f803ac0677"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.851096 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.851141 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sff2q\" (UniqueName: \"kubernetes.io/projected/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-kube-api-access-sff2q\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:45 crc kubenswrapper[4678]: I1206 11:10:45.851156 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d6e6e2e2-4745-46f6-a3fa-94f803ac0677-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.225919 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" event={"ID":"d6e6e2e2-4745-46f6-a3fa-94f803ac0677","Type":"ContainerDied","Data":"0590be8f2dd0918d1c795b32da9ded9231a5dbe787a69a5cc32499b22347b84d"} Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.225961 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0590be8f2dd0918d1c795b32da9ded9231a5dbe787a69a5cc32499b22347b84d" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.225971 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.313393 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt"] Dec 06 11:10:46 crc kubenswrapper[4678]: E1206 11:10:46.313870 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6e6e2e2-4745-46f6-a3fa-94f803ac0677" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.313894 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6e6e2e2-4745-46f6-a3fa-94f803ac0677" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.314111 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6e6e2e2-4745-46f6-a3fa-94f803ac0677" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.314766 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.317327 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.317336 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.317587 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.317902 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.318368 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.318824 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.319933 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.319985 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.328116 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt"] Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359144 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359215 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359245 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359291 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359315 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359464 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359531 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359601 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359642 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359698 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359738 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359774 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359825 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n59mz\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-kube-api-access-n59mz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.359899 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461628 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461682 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461733 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461776 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461803 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461928 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.461979 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.462007 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.462051 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.462090 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.462121 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.462150 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n59mz\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-kube-api-access-n59mz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.466964 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.467522 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.467595 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.467815 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.467896 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.468112 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.468809 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.469836 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.470249 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.471043 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.471302 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.471633 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.473501 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.481628 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n59mz\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-kube-api-access-n59mz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:46 crc kubenswrapper[4678]: I1206 11:10:46.633644 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:10:47 crc kubenswrapper[4678]: I1206 11:10:47.133260 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt"] Dec 06 11:10:47 crc kubenswrapper[4678]: I1206 11:10:47.234278 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" event={"ID":"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b","Type":"ContainerStarted","Data":"2c447a9eaa83297fa61ea7c820fd3614d1881f37d6a7bba280298135a2ba42d5"} Dec 06 11:10:48 crc kubenswrapper[4678]: I1206 11:10:48.246294 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" event={"ID":"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b","Type":"ContainerStarted","Data":"4fc19b66e68d79d217088b4b49befcdf2726be308ac63ac6913ebb721012dce6"} Dec 06 11:10:48 crc kubenswrapper[4678]: I1206 11:10:48.274754 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" podStartSLOduration=1.8634942429999999 podStartE2EDuration="2.274735956s" podCreationTimestamp="2025-12-06 11:10:46 +0000 UTC" firstStartedPulling="2025-12-06 11:10:47.142182901 +0000 UTC m=+2051.985614330" lastFinishedPulling="2025-12-06 11:10:47.553424584 +0000 UTC m=+2052.396856043" observedRunningTime="2025-12-06 11:10:48.26529945 +0000 UTC m=+2053.108730909" watchObservedRunningTime="2025-12-06 11:10:48.274735956 +0000 UTC m=+2053.118167405" Dec 06 11:10:59 crc kubenswrapper[4678]: I1206 11:10:59.505006 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:10:59 crc kubenswrapper[4678]: I1206 11:10:59.505616 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:11:29 crc kubenswrapper[4678]: I1206 11:11:29.505973 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:11:29 crc kubenswrapper[4678]: I1206 11:11:29.506485 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:11:29 crc kubenswrapper[4678]: I1206 11:11:29.506559 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:11:29 crc kubenswrapper[4678]: I1206 11:11:29.507027 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b5b29cf01b4f5ab8126b1632136a603737bbd297541220236e9787fb95d5818"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:11:29 crc kubenswrapper[4678]: I1206 11:11:29.507071 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://8b5b29cf01b4f5ab8126b1632136a603737bbd297541220236e9787fb95d5818" gracePeriod=600 Dec 06 11:11:30 crc kubenswrapper[4678]: I1206 11:11:30.615056 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="8b5b29cf01b4f5ab8126b1632136a603737bbd297541220236e9787fb95d5818" exitCode=0 Dec 06 11:11:30 crc kubenswrapper[4678]: I1206 11:11:30.615125 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"8b5b29cf01b4f5ab8126b1632136a603737bbd297541220236e9787fb95d5818"} Dec 06 11:11:30 crc kubenswrapper[4678]: I1206 11:11:30.615716 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5"} Dec 06 11:11:30 crc kubenswrapper[4678]: I1206 11:11:30.615770 4678 scope.go:117] "RemoveContainer" containerID="a701d9be17fe6ac402fcceae15b18358e28f1c3293971084e705a1ac290b55da" Dec 06 11:11:32 crc kubenswrapper[4678]: I1206 11:11:32.642292 4678 generic.go:334] "Generic (PLEG): container finished" podID="a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" containerID="4fc19b66e68d79d217088b4b49befcdf2726be308ac63ac6913ebb721012dce6" exitCode=0 Dec 06 11:11:32 crc kubenswrapper[4678]: I1206 11:11:32.642591 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" event={"ID":"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b","Type":"ContainerDied","Data":"4fc19b66e68d79d217088b4b49befcdf2726be308ac63ac6913ebb721012dce6"} Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.067673 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198388 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-bootstrap-combined-ca-bundle\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198451 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-nova-combined-ca-bundle\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198520 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198598 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ssh-key\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198685 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-telemetry-combined-ca-bundle\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198709 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ovn-combined-ca-bundle\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198727 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-repo-setup-combined-ca-bundle\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198745 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198778 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-neutron-metadata-combined-ca-bundle\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198802 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-libvirt-combined-ca-bundle\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198835 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198857 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n59mz\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-kube-api-access-n59mz\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198873 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-inventory\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.198904 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-ovn-default-certs-0\") pod \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\" (UID: \"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b\") " Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.206381 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.206637 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.206968 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.207080 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.208123 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.208797 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.209446 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.212760 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.212804 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.212953 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-kube-api-access-n59mz" (OuterVolumeSpecName: "kube-api-access-n59mz") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "kube-api-access-n59mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.213072 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.213750 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.241559 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.259805 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-inventory" (OuterVolumeSpecName: "inventory") pod "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" (UID: "a8f9dc0d-5e17-4e1b-9363-2d193c592e9b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.300864 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301125 4678 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301199 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301263 4678 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301330 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301422 4678 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301522 4678 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301605 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301667 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n59mz\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-kube-api-access-n59mz\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301757 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301816 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301876 4678 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.301937 4678 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.302001 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a8f9dc0d-5e17-4e1b-9363-2d193c592e9b-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.663859 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" event={"ID":"a8f9dc0d-5e17-4e1b-9363-2d193c592e9b","Type":"ContainerDied","Data":"2c447a9eaa83297fa61ea7c820fd3614d1881f37d6a7bba280298135a2ba42d5"} Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.664068 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c447a9eaa83297fa61ea7c820fd3614d1881f37d6a7bba280298135a2ba42d5" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.663923 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.799183 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9"] Dec 06 11:11:34 crc kubenswrapper[4678]: E1206 11:11:34.799588 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.799606 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.799770 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8f9dc0d-5e17-4e1b-9363-2d193c592e9b" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.800342 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.804202 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.804442 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.804592 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.805073 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.805244 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.808463 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9"] Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.914364 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.914407 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.914457 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.914519 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c79d4229-477e-4f22-90d8-449984bab4cb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:34 crc kubenswrapper[4678]: I1206 11:11:34.914564 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qwwq\" (UniqueName: \"kubernetes.io/projected/c79d4229-477e-4f22-90d8-449984bab4cb-kube-api-access-2qwwq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.015823 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.015881 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.015936 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.015999 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c79d4229-477e-4f22-90d8-449984bab4cb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.016064 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qwwq\" (UniqueName: \"kubernetes.io/projected/c79d4229-477e-4f22-90d8-449984bab4cb-kube-api-access-2qwwq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.017245 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c79d4229-477e-4f22-90d8-449984bab4cb-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.021367 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.023466 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.030511 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.037328 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qwwq\" (UniqueName: \"kubernetes.io/projected/c79d4229-477e-4f22-90d8-449984bab4cb-kube-api-access-2qwwq\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-7xqv9\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.120276 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:11:35 crc kubenswrapper[4678]: I1206 11:11:35.663677 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9"] Dec 06 11:11:36 crc kubenswrapper[4678]: I1206 11:11:36.423264 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:11:36 crc kubenswrapper[4678]: I1206 11:11:36.687027 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" event={"ID":"c79d4229-477e-4f22-90d8-449984bab4cb","Type":"ContainerStarted","Data":"1fac3a6257b31914f7e82319aef56d2ef2a625d9790961b020d26d412825f1ea"} Dec 06 11:11:37 crc kubenswrapper[4678]: I1206 11:11:37.699029 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" event={"ID":"c79d4229-477e-4f22-90d8-449984bab4cb","Type":"ContainerStarted","Data":"9227a72a4832d75d58f93e3937c16ddfcf49a0f3bf44d42058df767b070416eb"} Dec 06 11:11:37 crc kubenswrapper[4678]: I1206 11:11:37.723893 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" podStartSLOduration=2.981863589 podStartE2EDuration="3.723875903s" podCreationTimestamp="2025-12-06 11:11:34 +0000 UTC" firstStartedPulling="2025-12-06 11:11:35.678339382 +0000 UTC m=+2100.521770821" lastFinishedPulling="2025-12-06 11:11:36.420351686 +0000 UTC m=+2101.263783135" observedRunningTime="2025-12-06 11:11:37.714765406 +0000 UTC m=+2102.558196845" watchObservedRunningTime="2025-12-06 11:11:37.723875903 +0000 UTC m=+2102.567307342" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.622379 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-f9lq4"] Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.625317 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.691089 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f9lq4"] Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.773897 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-utilities\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.773969 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/25454492-ba53-4cc9-865a-ab6383ee9c54-kube-api-access-f54zq\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.774039 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-catalog-content\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.876042 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-utilities\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.876091 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/25454492-ba53-4cc9-865a-ab6383ee9c54-kube-api-access-f54zq\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.876126 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-catalog-content\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.876704 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-catalog-content\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.877115 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-utilities\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.895771 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/25454492-ba53-4cc9-865a-ab6383ee9c54-kube-api-access-f54zq\") pod \"certified-operators-f9lq4\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:49 crc kubenswrapper[4678]: I1206 11:12:49.945368 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:50 crc kubenswrapper[4678]: I1206 11:12:50.512358 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-f9lq4"] Dec 06 11:12:51 crc kubenswrapper[4678]: I1206 11:12:51.482960 4678 generic.go:334] "Generic (PLEG): container finished" podID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerID="9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e" exitCode=0 Dec 06 11:12:51 crc kubenswrapper[4678]: I1206 11:12:51.486987 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f9lq4" event={"ID":"25454492-ba53-4cc9-865a-ab6383ee9c54","Type":"ContainerDied","Data":"9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e"} Dec 06 11:12:51 crc kubenswrapper[4678]: I1206 11:12:51.487046 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f9lq4" event={"ID":"25454492-ba53-4cc9-865a-ab6383ee9c54","Type":"ContainerStarted","Data":"ae11f70bb0fe09768f77ca56fe110d807303391eb983ca4e9e6078a0e27e9340"} Dec 06 11:12:52 crc kubenswrapper[4678]: I1206 11:12:52.492428 4678 generic.go:334] "Generic (PLEG): container finished" podID="c79d4229-477e-4f22-90d8-449984bab4cb" containerID="9227a72a4832d75d58f93e3937c16ddfcf49a0f3bf44d42058df767b070416eb" exitCode=0 Dec 06 11:12:52 crc kubenswrapper[4678]: I1206 11:12:52.492532 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" event={"ID":"c79d4229-477e-4f22-90d8-449984bab4cb","Type":"ContainerDied","Data":"9227a72a4832d75d58f93e3937c16ddfcf49a0f3bf44d42058df767b070416eb"} Dec 06 11:12:52 crc kubenswrapper[4678]: I1206 11:12:52.494530 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f9lq4" event={"ID":"25454492-ba53-4cc9-865a-ab6383ee9c54","Type":"ContainerStarted","Data":"0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb"} Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.045927 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.174539 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ovn-combined-ca-bundle\") pod \"c79d4229-477e-4f22-90d8-449984bab4cb\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.174592 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ssh-key\") pod \"c79d4229-477e-4f22-90d8-449984bab4cb\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.174625 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c79d4229-477e-4f22-90d8-449984bab4cb-ovncontroller-config-0\") pod \"c79d4229-477e-4f22-90d8-449984bab4cb\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.174781 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-inventory\") pod \"c79d4229-477e-4f22-90d8-449984bab4cb\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.174854 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qwwq\" (UniqueName: \"kubernetes.io/projected/c79d4229-477e-4f22-90d8-449984bab4cb-kube-api-access-2qwwq\") pod \"c79d4229-477e-4f22-90d8-449984bab4cb\" (UID: \"c79d4229-477e-4f22-90d8-449984bab4cb\") " Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.191051 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c79d4229-477e-4f22-90d8-449984bab4cb" (UID: "c79d4229-477e-4f22-90d8-449984bab4cb"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.193182 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c79d4229-477e-4f22-90d8-449984bab4cb-kube-api-access-2qwwq" (OuterVolumeSpecName: "kube-api-access-2qwwq") pod "c79d4229-477e-4f22-90d8-449984bab4cb" (UID: "c79d4229-477e-4f22-90d8-449984bab4cb"). InnerVolumeSpecName "kube-api-access-2qwwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.208695 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-inventory" (OuterVolumeSpecName: "inventory") pod "c79d4229-477e-4f22-90d8-449984bab4cb" (UID: "c79d4229-477e-4f22-90d8-449984bab4cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.215764 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c79d4229-477e-4f22-90d8-449984bab4cb-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "c79d4229-477e-4f22-90d8-449984bab4cb" (UID: "c79d4229-477e-4f22-90d8-449984bab4cb"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.231873 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c79d4229-477e-4f22-90d8-449984bab4cb" (UID: "c79d4229-477e-4f22-90d8-449984bab4cb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.279574 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.279608 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qwwq\" (UniqueName: \"kubernetes.io/projected/c79d4229-477e-4f22-90d8-449984bab4cb-kube-api-access-2qwwq\") on node \"crc\" DevicePath \"\"" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.279619 4678 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.279627 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c79d4229-477e-4f22-90d8-449984bab4cb-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.279637 4678 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/c79d4229-477e-4f22-90d8-449984bab4cb-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.520914 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" event={"ID":"c79d4229-477e-4f22-90d8-449984bab4cb","Type":"ContainerDied","Data":"1fac3a6257b31914f7e82319aef56d2ef2a625d9790961b020d26d412825f1ea"} Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.521328 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fac3a6257b31914f7e82319aef56d2ef2a625d9790961b020d26d412825f1ea" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.521453 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-7xqv9" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.524340 4678 generic.go:334] "Generic (PLEG): container finished" podID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerID="0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb" exitCode=0 Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.524396 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f9lq4" event={"ID":"25454492-ba53-4cc9-865a-ab6383ee9c54","Type":"ContainerDied","Data":"0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb"} Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.731841 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6"] Dec 06 11:12:54 crc kubenswrapper[4678]: E1206 11:12:54.732273 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c79d4229-477e-4f22-90d8-449984bab4cb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.732294 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c79d4229-477e-4f22-90d8-449984bab4cb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.732533 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c79d4229-477e-4f22-90d8-449984bab4cb" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.733165 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.735781 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.736422 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.736481 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.736978 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.737067 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.743957 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.762018 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6"] Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.891480 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.891775 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.891884 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.891947 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.891990 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6g88\" (UniqueName: \"kubernetes.io/projected/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-kube-api-access-g6g88\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.892010 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.996099 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.996155 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6g88\" (UniqueName: \"kubernetes.io/projected/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-kube-api-access-g6g88\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.996180 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.996306 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.996346 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:54 crc kubenswrapper[4678]: I1206 11:12:54.996384 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.001227 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.003827 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.009338 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.012011 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.027336 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.028289 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6g88\" (UniqueName: \"kubernetes.io/projected/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-kube-api-access-g6g88\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.051960 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.536322 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f9lq4" event={"ID":"25454492-ba53-4cc9-865a-ab6383ee9c54","Type":"ContainerStarted","Data":"1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171"} Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.561765 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-f9lq4" podStartSLOduration=3.093995337 podStartE2EDuration="6.56175161s" podCreationTimestamp="2025-12-06 11:12:49 +0000 UTC" firstStartedPulling="2025-12-06 11:12:51.484603031 +0000 UTC m=+2176.328034470" lastFinishedPulling="2025-12-06 11:12:54.952359304 +0000 UTC m=+2179.795790743" observedRunningTime="2025-12-06 11:12:55.559909541 +0000 UTC m=+2180.403340980" watchObservedRunningTime="2025-12-06 11:12:55.56175161 +0000 UTC m=+2180.405183049" Dec 06 11:12:55 crc kubenswrapper[4678]: I1206 11:12:55.640457 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6"] Dec 06 11:12:55 crc kubenswrapper[4678]: W1206 11:12:55.641795 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e7abea2_75f9_4bcb_897d_d62fe1e7be3e.slice/crio-ce0f3586e05c1e55b3bd688c47d6265aca5da9d27179020fe9595b03ca5433e9 WatchSource:0}: Error finding container ce0f3586e05c1e55b3bd688c47d6265aca5da9d27179020fe9595b03ca5433e9: Status 404 returned error can't find the container with id ce0f3586e05c1e55b3bd688c47d6265aca5da9d27179020fe9595b03ca5433e9 Dec 06 11:12:56 crc kubenswrapper[4678]: I1206 11:12:56.546298 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" event={"ID":"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e","Type":"ContainerStarted","Data":"83bb34a1888a08de36ea6f1198311e2d66de8d69f18e870263350b660da17baf"} Dec 06 11:12:56 crc kubenswrapper[4678]: I1206 11:12:56.546681 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" event={"ID":"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e","Type":"ContainerStarted","Data":"ce0f3586e05c1e55b3bd688c47d6265aca5da9d27179020fe9595b03ca5433e9"} Dec 06 11:12:56 crc kubenswrapper[4678]: I1206 11:12:56.566582 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" podStartSLOduration=2.026648718 podStartE2EDuration="2.566565023s" podCreationTimestamp="2025-12-06 11:12:54 +0000 UTC" firstStartedPulling="2025-12-06 11:12:55.643459806 +0000 UTC m=+2180.486891245" lastFinishedPulling="2025-12-06 11:12:56.183376101 +0000 UTC m=+2181.026807550" observedRunningTime="2025-12-06 11:12:56.559273723 +0000 UTC m=+2181.402705162" watchObservedRunningTime="2025-12-06 11:12:56.566565023 +0000 UTC m=+2181.409996462" Dec 06 11:12:59 crc kubenswrapper[4678]: I1206 11:12:59.946033 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:59 crc kubenswrapper[4678]: I1206 11:12:59.946760 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:12:59 crc kubenswrapper[4678]: I1206 11:12:59.997380 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:13:00 crc kubenswrapper[4678]: I1206 11:13:00.630795 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:13:00 crc kubenswrapper[4678]: I1206 11:13:00.685455 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f9lq4"] Dec 06 11:13:02 crc kubenswrapper[4678]: I1206 11:13:02.599921 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-f9lq4" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="registry-server" containerID="cri-o://1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171" gracePeriod=2 Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.076471 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.195138 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/25454492-ba53-4cc9-865a-ab6383ee9c54-kube-api-access-f54zq\") pod \"25454492-ba53-4cc9-865a-ab6383ee9c54\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.195210 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-utilities\") pod \"25454492-ba53-4cc9-865a-ab6383ee9c54\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.195241 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-catalog-content\") pod \"25454492-ba53-4cc9-865a-ab6383ee9c54\" (UID: \"25454492-ba53-4cc9-865a-ab6383ee9c54\") " Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.196082 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-utilities" (OuterVolumeSpecName: "utilities") pod "25454492-ba53-4cc9-865a-ab6383ee9c54" (UID: "25454492-ba53-4cc9-865a-ab6383ee9c54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.211314 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25454492-ba53-4cc9-865a-ab6383ee9c54-kube-api-access-f54zq" (OuterVolumeSpecName: "kube-api-access-f54zq") pod "25454492-ba53-4cc9-865a-ab6383ee9c54" (UID: "25454492-ba53-4cc9-865a-ab6383ee9c54"). InnerVolumeSpecName "kube-api-access-f54zq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.243076 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25454492-ba53-4cc9-865a-ab6383ee9c54" (UID: "25454492-ba53-4cc9-865a-ab6383ee9c54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.298008 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f54zq\" (UniqueName: \"kubernetes.io/projected/25454492-ba53-4cc9-865a-ab6383ee9c54-kube-api-access-f54zq\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.298043 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.298057 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25454492-ba53-4cc9-865a-ab6383ee9c54-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.610374 4678 generic.go:334] "Generic (PLEG): container finished" podID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerID="1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171" exitCode=0 Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.610531 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-f9lq4" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.610556 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f9lq4" event={"ID":"25454492-ba53-4cc9-865a-ab6383ee9c54","Type":"ContainerDied","Data":"1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171"} Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.611372 4678 scope.go:117] "RemoveContainer" containerID="1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.612036 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-f9lq4" event={"ID":"25454492-ba53-4cc9-865a-ab6383ee9c54","Type":"ContainerDied","Data":"ae11f70bb0fe09768f77ca56fe110d807303391eb983ca4e9e6078a0e27e9340"} Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.641307 4678 scope.go:117] "RemoveContainer" containerID="0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.646820 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-f9lq4"] Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.663879 4678 scope.go:117] "RemoveContainer" containerID="9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.670182 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-f9lq4"] Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.731729 4678 scope.go:117] "RemoveContainer" containerID="1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171" Dec 06 11:13:03 crc kubenswrapper[4678]: E1206 11:13:03.736075 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171\": container with ID starting with 1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171 not found: ID does not exist" containerID="1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.736142 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171"} err="failed to get container status \"1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171\": rpc error: code = NotFound desc = could not find container \"1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171\": container with ID starting with 1386cddd529ade72669b198ffc9ab21c2b95829b56351bce3042553da540e171 not found: ID does not exist" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.736197 4678 scope.go:117] "RemoveContainer" containerID="0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb" Dec 06 11:13:03 crc kubenswrapper[4678]: E1206 11:13:03.737144 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb\": container with ID starting with 0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb not found: ID does not exist" containerID="0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.737175 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb"} err="failed to get container status \"0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb\": rpc error: code = NotFound desc = could not find container \"0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb\": container with ID starting with 0f8ab061d448bc3d7cf6ee2a627ddf7e5b6f73e575b2b530e567a50911b4e7fb not found: ID does not exist" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.737293 4678 scope.go:117] "RemoveContainer" containerID="9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e" Dec 06 11:13:03 crc kubenswrapper[4678]: E1206 11:13:03.738082 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e\": container with ID starting with 9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e not found: ID does not exist" containerID="9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e" Dec 06 11:13:03 crc kubenswrapper[4678]: I1206 11:13:03.738108 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e"} err="failed to get container status \"9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e\": rpc error: code = NotFound desc = could not find container \"9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e\": container with ID starting with 9380a42780e7f75db40ca613ef7c47cc71c380b2adc50675fa6cf85534772b1e not found: ID does not exist" Dec 06 11:13:03 crc kubenswrapper[4678]: E1206 11:13:03.790118 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25454492_ba53_4cc9_865a_ab6383ee9c54.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25454492_ba53_4cc9_865a_ab6383ee9c54.slice/crio-ae11f70bb0fe09768f77ca56fe110d807303391eb983ca4e9e6078a0e27e9340\": RecentStats: unable to find data in memory cache]" Dec 06 11:13:05 crc kubenswrapper[4678]: I1206 11:13:05.493890 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" path="/var/lib/kubelet/pods/25454492-ba53-4cc9-865a-ab6383ee9c54/volumes" Dec 06 11:13:29 crc kubenswrapper[4678]: I1206 11:13:29.505371 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:13:29 crc kubenswrapper[4678]: I1206 11:13:29.505945 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.081674 4678 generic.go:334] "Generic (PLEG): container finished" podID="1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" containerID="83bb34a1888a08de36ea6f1198311e2d66de8d69f18e870263350b660da17baf" exitCode=0 Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.081767 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" event={"ID":"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e","Type":"ContainerDied","Data":"83bb34a1888a08de36ea6f1198311e2d66de8d69f18e870263350b660da17baf"} Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.582336 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lc89v"] Dec 06 11:13:52 crc kubenswrapper[4678]: E1206 11:13:52.583249 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="extract-content" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.583270 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="extract-content" Dec 06 11:13:52 crc kubenswrapper[4678]: E1206 11:13:52.583284 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="registry-server" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.583291 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="registry-server" Dec 06 11:13:52 crc kubenswrapper[4678]: E1206 11:13:52.583313 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="extract-utilities" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.583320 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="extract-utilities" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.583634 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="25454492-ba53-4cc9-865a-ab6383ee9c54" containerName="registry-server" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.585131 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.593212 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-catalog-content\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.593424 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-utilities\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.593553 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7f2n\" (UniqueName: \"kubernetes.io/projected/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-kube-api-access-g7f2n\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.604540 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lc89v"] Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.697566 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-catalog-content\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.697820 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-utilities\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.697950 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7f2n\" (UniqueName: \"kubernetes.io/projected/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-kube-api-access-g7f2n\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.698066 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-catalog-content\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.698577 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-utilities\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.727247 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7f2n\" (UniqueName: \"kubernetes.io/projected/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-kube-api-access-g7f2n\") pod \"redhat-marketplace-lc89v\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:52 crc kubenswrapper[4678]: I1206 11:13:52.902574 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.443595 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lc89v"] Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.609773 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.718034 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-nova-metadata-neutron-config-0\") pod \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.718094 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-ssh-key\") pod \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.718113 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-metadata-combined-ca-bundle\") pod \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.718136 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6g88\" (UniqueName: \"kubernetes.io/projected/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-kube-api-access-g6g88\") pod \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.718178 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-inventory\") pod \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.718252 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-ovn-metadata-agent-neutron-config-0\") pod \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\" (UID: \"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e\") " Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.725560 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-kube-api-access-g6g88" (OuterVolumeSpecName: "kube-api-access-g6g88") pod "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" (UID: "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e"). InnerVolumeSpecName "kube-api-access-g6g88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.745050 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" (UID: "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.747727 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" (UID: "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.762301 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" (UID: "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.764661 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" (UID: "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.776666 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-inventory" (OuterVolumeSpecName: "inventory") pod "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" (UID: "1e7abea2-75f9-4bcb-897d-d62fe1e7be3e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.820070 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.820097 4678 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.820110 4678 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.820122 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.820132 4678 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:53 crc kubenswrapper[4678]: I1206 11:13:53.820142 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6g88\" (UniqueName: \"kubernetes.io/projected/1e7abea2-75f9-4bcb-897d-d62fe1e7be3e-kube-api-access-g6g88\") on node \"crc\" DevicePath \"\"" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.100772 4678 generic.go:334] "Generic (PLEG): container finished" podID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerID="432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83" exitCode=0 Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.100838 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lc89v" event={"ID":"3b7f1288-f94d-4777-82dd-7e45f38fc2f3","Type":"ContainerDied","Data":"432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83"} Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.100872 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lc89v" event={"ID":"3b7f1288-f94d-4777-82dd-7e45f38fc2f3","Type":"ContainerStarted","Data":"b3dd270f8b1482fd2eeb93a8c0888c3cd8f8e43cb0c6178b2351f23719db5f3c"} Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.103452 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" event={"ID":"1e7abea2-75f9-4bcb-897d-d62fe1e7be3e","Type":"ContainerDied","Data":"ce0f3586e05c1e55b3bd688c47d6265aca5da9d27179020fe9595b03ca5433e9"} Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.103590 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce0f3586e05c1e55b3bd688c47d6265aca5da9d27179020fe9595b03ca5433e9" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.103510 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.194899 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl"] Dec 06 11:13:54 crc kubenswrapper[4678]: E1206 11:13:54.195414 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.195476 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.195721 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e7abea2-75f9-4bcb-897d-d62fe1e7be3e" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.196323 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.198248 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.199418 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.201606 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.202353 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.208173 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl"] Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.210666 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.360217 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.360472 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.360509 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.360576 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.361175 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6ggl\" (UniqueName: \"kubernetes.io/projected/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-kube-api-access-l6ggl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.462659 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.462737 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.462765 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.462832 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.462906 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6ggl\" (UniqueName: \"kubernetes.io/projected/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-kube-api-access-l6ggl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.466302 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.466324 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.467017 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.466781 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.479711 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6ggl\" (UniqueName: \"kubernetes.io/projected/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-kube-api-access-l6ggl\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.513300 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.768168 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kjtvm"] Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.771126 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.776159 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kjtvm"] Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.872321 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-utilities\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.872385 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-catalog-content\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.872424 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tgn6\" (UniqueName: \"kubernetes.io/projected/8e30186e-a4f8-4480-90e3-cfa66d3b200b-kube-api-access-2tgn6\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.972530 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9vznd"] Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.973705 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-catalog-content\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.973832 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tgn6\" (UniqueName: \"kubernetes.io/projected/8e30186e-a4f8-4480-90e3-cfa66d3b200b-kube-api-access-2tgn6\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.974001 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-utilities\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.974518 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-utilities\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.974925 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-catalog-content\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.975450 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:54 crc kubenswrapper[4678]: I1206 11:13:54.984102 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9vznd"] Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.003376 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tgn6\" (UniqueName: \"kubernetes.io/projected/8e30186e-a4f8-4480-90e3-cfa66d3b200b-kube-api-access-2tgn6\") pod \"community-operators-kjtvm\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.117763 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-catalog-content\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.118113 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rq49\" (UniqueName: \"kubernetes.io/projected/c976648c-0d90-42c1-b2ed-5071fa35b85d-kube-api-access-7rq49\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.118329 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-utilities\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.119139 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl"] Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.140039 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" event={"ID":"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8","Type":"ContainerStarted","Data":"15bf7a0fa6013d68fdbed79482db2841ded77970f28c246809f66957b5c910d0"} Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.142293 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lc89v" event={"ID":"3b7f1288-f94d-4777-82dd-7e45f38fc2f3","Type":"ContainerStarted","Data":"36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff"} Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.166659 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.221628 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-catalog-content\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.221782 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rq49\" (UniqueName: \"kubernetes.io/projected/c976648c-0d90-42c1-b2ed-5071fa35b85d-kube-api-access-7rq49\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.221971 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-utilities\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.223243 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-utilities\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.224711 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-catalog-content\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.260339 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rq49\" (UniqueName: \"kubernetes.io/projected/c976648c-0d90-42c1-b2ed-5071fa35b85d-kube-api-access-7rq49\") pod \"redhat-operators-9vznd\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.301716 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:13:55 crc kubenswrapper[4678]: I1206 11:13:55.881406 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kjtvm"] Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.039138 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9vznd"] Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.151230 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" event={"ID":"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8","Type":"ContainerStarted","Data":"c37d25a0f03941ed056beedfc2d37cf8f7571bb135eec3b7656777d7c5e5b3f3"} Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.153799 4678 generic.go:334] "Generic (PLEG): container finished" podID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerID="36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff" exitCode=0 Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.154008 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lc89v" event={"ID":"3b7f1288-f94d-4777-82dd-7e45f38fc2f3","Type":"ContainerDied","Data":"36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff"} Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.155371 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vznd" event={"ID":"c976648c-0d90-42c1-b2ed-5071fa35b85d","Type":"ContainerStarted","Data":"fab178dabb0e9f29c904bba0aa80f8bb389afdf52ed9f21da6aba041edd36d1f"} Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.166027 4678 generic.go:334] "Generic (PLEG): container finished" podID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerID="203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631" exitCode=0 Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.166072 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjtvm" event={"ID":"8e30186e-a4f8-4480-90e3-cfa66d3b200b","Type":"ContainerDied","Data":"203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631"} Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.166102 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjtvm" event={"ID":"8e30186e-a4f8-4480-90e3-cfa66d3b200b","Type":"ContainerStarted","Data":"81d8ddfd27d1357a6eca4da12c17db26fa89f6ae99df09df278a3a62b0e6696c"} Dec 06 11:13:56 crc kubenswrapper[4678]: I1206 11:13:56.174771 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" podStartSLOduration=1.731215185 podStartE2EDuration="2.174755771s" podCreationTimestamp="2025-12-06 11:13:54 +0000 UTC" firstStartedPulling="2025-12-06 11:13:55.090686918 +0000 UTC m=+2239.934118357" lastFinishedPulling="2025-12-06 11:13:55.534227504 +0000 UTC m=+2240.377658943" observedRunningTime="2025-12-06 11:13:56.169406312 +0000 UTC m=+2241.012837751" watchObservedRunningTime="2025-12-06 11:13:56.174755771 +0000 UTC m=+2241.018187210" Dec 06 11:13:57 crc kubenswrapper[4678]: I1206 11:13:57.188002 4678 generic.go:334] "Generic (PLEG): container finished" podID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerID="f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62" exitCode=0 Dec 06 11:13:57 crc kubenswrapper[4678]: I1206 11:13:57.188315 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vznd" event={"ID":"c976648c-0d90-42c1-b2ed-5071fa35b85d","Type":"ContainerDied","Data":"f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62"} Dec 06 11:13:57 crc kubenswrapper[4678]: I1206 11:13:57.193582 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjtvm" event={"ID":"8e30186e-a4f8-4480-90e3-cfa66d3b200b","Type":"ContainerStarted","Data":"fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa"} Dec 06 11:13:57 crc kubenswrapper[4678]: I1206 11:13:57.197365 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lc89v" event={"ID":"3b7f1288-f94d-4777-82dd-7e45f38fc2f3","Type":"ContainerStarted","Data":"90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007"} Dec 06 11:13:57 crc kubenswrapper[4678]: I1206 11:13:57.232946 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lc89v" podStartSLOduration=2.798664862 podStartE2EDuration="5.232931957s" podCreationTimestamp="2025-12-06 11:13:52 +0000 UTC" firstStartedPulling="2025-12-06 11:13:54.103082748 +0000 UTC m=+2238.946514217" lastFinishedPulling="2025-12-06 11:13:56.537349873 +0000 UTC m=+2241.380781312" observedRunningTime="2025-12-06 11:13:57.22575761 +0000 UTC m=+2242.069189059" watchObservedRunningTime="2025-12-06 11:13:57.232931957 +0000 UTC m=+2242.076363396" Dec 06 11:13:58 crc kubenswrapper[4678]: I1206 11:13:58.209695 4678 generic.go:334] "Generic (PLEG): container finished" podID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerID="fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa" exitCode=0 Dec 06 11:13:58 crc kubenswrapper[4678]: I1206 11:13:58.209791 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjtvm" event={"ID":"8e30186e-a4f8-4480-90e3-cfa66d3b200b","Type":"ContainerDied","Data":"fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa"} Dec 06 11:13:59 crc kubenswrapper[4678]: I1206 11:13:59.225758 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vznd" event={"ID":"c976648c-0d90-42c1-b2ed-5071fa35b85d","Type":"ContainerStarted","Data":"e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a"} Dec 06 11:13:59 crc kubenswrapper[4678]: I1206 11:13:59.506407 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:13:59 crc kubenswrapper[4678]: I1206 11:13:59.506476 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:14:00 crc kubenswrapper[4678]: I1206 11:14:00.253148 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjtvm" event={"ID":"8e30186e-a4f8-4480-90e3-cfa66d3b200b","Type":"ContainerStarted","Data":"8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b"} Dec 06 11:14:00 crc kubenswrapper[4678]: I1206 11:14:00.311938 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kjtvm" podStartSLOduration=3.590584903 podStartE2EDuration="6.311918861s" podCreationTimestamp="2025-12-06 11:13:54 +0000 UTC" firstStartedPulling="2025-12-06 11:13:56.168569125 +0000 UTC m=+2241.012000564" lastFinishedPulling="2025-12-06 11:13:58.889903073 +0000 UTC m=+2243.733334522" observedRunningTime="2025-12-06 11:14:00.30521117 +0000 UTC m=+2245.148642609" watchObservedRunningTime="2025-12-06 11:14:00.311918861 +0000 UTC m=+2245.155350300" Dec 06 11:14:01 crc kubenswrapper[4678]: I1206 11:14:01.265003 4678 generic.go:334] "Generic (PLEG): container finished" podID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerID="e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a" exitCode=0 Dec 06 11:14:01 crc kubenswrapper[4678]: I1206 11:14:01.265007 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vznd" event={"ID":"c976648c-0d90-42c1-b2ed-5071fa35b85d","Type":"ContainerDied","Data":"e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a"} Dec 06 11:14:02 crc kubenswrapper[4678]: I1206 11:14:02.276265 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vznd" event={"ID":"c976648c-0d90-42c1-b2ed-5071fa35b85d","Type":"ContainerStarted","Data":"3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064"} Dec 06 11:14:02 crc kubenswrapper[4678]: I1206 11:14:02.292812 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9vznd" podStartSLOduration=3.806516283 podStartE2EDuration="8.292796312s" podCreationTimestamp="2025-12-06 11:13:54 +0000 UTC" firstStartedPulling="2025-12-06 11:13:57.189717714 +0000 UTC m=+2242.033149153" lastFinishedPulling="2025-12-06 11:14:01.675997733 +0000 UTC m=+2246.519429182" observedRunningTime="2025-12-06 11:14:02.292206123 +0000 UTC m=+2247.135637562" watchObservedRunningTime="2025-12-06 11:14:02.292796312 +0000 UTC m=+2247.136227751" Dec 06 11:14:02 crc kubenswrapper[4678]: I1206 11:14:02.903692 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:14:02 crc kubenswrapper[4678]: I1206 11:14:02.903751 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:14:02 crc kubenswrapper[4678]: I1206 11:14:02.949836 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:14:03 crc kubenswrapper[4678]: I1206 11:14:03.339928 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:14:05 crc kubenswrapper[4678]: I1206 11:14:05.167301 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:14:05 crc kubenswrapper[4678]: I1206 11:14:05.167554 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:14:05 crc kubenswrapper[4678]: I1206 11:14:05.303452 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:14:05 crc kubenswrapper[4678]: I1206 11:14:05.303509 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:14:06 crc kubenswrapper[4678]: I1206 11:14:06.213971 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-kjtvm" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="registry-server" probeResult="failure" output=< Dec 06 11:14:06 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:14:06 crc kubenswrapper[4678]: > Dec 06 11:14:06 crc kubenswrapper[4678]: I1206 11:14:06.348619 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9vznd" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="registry-server" probeResult="failure" output=< Dec 06 11:14:06 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:14:06 crc kubenswrapper[4678]: > Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.361679 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lc89v"] Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.362039 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lc89v" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="registry-server" containerID="cri-o://90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007" gracePeriod=2 Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.786506 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.967641 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-catalog-content\") pod \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.967776 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-utilities\") pod \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.967891 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7f2n\" (UniqueName: \"kubernetes.io/projected/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-kube-api-access-g7f2n\") pod \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\" (UID: \"3b7f1288-f94d-4777-82dd-7e45f38fc2f3\") " Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.969538 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-utilities" (OuterVolumeSpecName: "utilities") pod "3b7f1288-f94d-4777-82dd-7e45f38fc2f3" (UID: "3b7f1288-f94d-4777-82dd-7e45f38fc2f3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:14:07 crc kubenswrapper[4678]: I1206 11:14:07.975760 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-kube-api-access-g7f2n" (OuterVolumeSpecName: "kube-api-access-g7f2n") pod "3b7f1288-f94d-4777-82dd-7e45f38fc2f3" (UID: "3b7f1288-f94d-4777-82dd-7e45f38fc2f3"). InnerVolumeSpecName "kube-api-access-g7f2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.006628 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b7f1288-f94d-4777-82dd-7e45f38fc2f3" (UID: "3b7f1288-f94d-4777-82dd-7e45f38fc2f3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.069695 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.069730 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7f2n\" (UniqueName: \"kubernetes.io/projected/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-kube-api-access-g7f2n\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.069742 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b7f1288-f94d-4777-82dd-7e45f38fc2f3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.329998 4678 generic.go:334] "Generic (PLEG): container finished" podID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerID="90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007" exitCode=0 Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.330047 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lc89v" event={"ID":"3b7f1288-f94d-4777-82dd-7e45f38fc2f3","Type":"ContainerDied","Data":"90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007"} Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.330078 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lc89v" event={"ID":"3b7f1288-f94d-4777-82dd-7e45f38fc2f3","Type":"ContainerDied","Data":"b3dd270f8b1482fd2eeb93a8c0888c3cd8f8e43cb0c6178b2351f23719db5f3c"} Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.330103 4678 scope.go:117] "RemoveContainer" containerID="90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.330267 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lc89v" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.370771 4678 scope.go:117] "RemoveContainer" containerID="36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.370811 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lc89v"] Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.378692 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lc89v"] Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.406356 4678 scope.go:117] "RemoveContainer" containerID="432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.443539 4678 scope.go:117] "RemoveContainer" containerID="90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007" Dec 06 11:14:08 crc kubenswrapper[4678]: E1206 11:14:08.444150 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007\": container with ID starting with 90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007 not found: ID does not exist" containerID="90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.444191 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007"} err="failed to get container status \"90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007\": rpc error: code = NotFound desc = could not find container \"90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007\": container with ID starting with 90c7fb3e0333ade5f094ac5603bd2c35ecbaedd15b8a89046ce5691adbcbe007 not found: ID does not exist" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.444216 4678 scope.go:117] "RemoveContainer" containerID="36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff" Dec 06 11:14:08 crc kubenswrapper[4678]: E1206 11:14:08.444709 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff\": container with ID starting with 36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff not found: ID does not exist" containerID="36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.444732 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff"} err="failed to get container status \"36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff\": rpc error: code = NotFound desc = could not find container \"36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff\": container with ID starting with 36c8b1b8f17fc8d86508bf287b55c4f9af49b938ce010c8299a7c14c0003b0ff not found: ID does not exist" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.444746 4678 scope.go:117] "RemoveContainer" containerID="432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83" Dec 06 11:14:08 crc kubenswrapper[4678]: E1206 11:14:08.444970 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83\": container with ID starting with 432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83 not found: ID does not exist" containerID="432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83" Dec 06 11:14:08 crc kubenswrapper[4678]: I1206 11:14:08.444988 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83"} err="failed to get container status \"432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83\": rpc error: code = NotFound desc = could not find container \"432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83\": container with ID starting with 432f28a25596ebd34d2bfc67203f592443c6ddbee2dd6cc3b4b9190dad8efa83 not found: ID does not exist" Dec 06 11:14:09 crc kubenswrapper[4678]: I1206 11:14:09.492340 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" path="/var/lib/kubelet/pods/3b7f1288-f94d-4777-82dd-7e45f38fc2f3/volumes" Dec 06 11:14:15 crc kubenswrapper[4678]: I1206 11:14:15.213800 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:14:15 crc kubenswrapper[4678]: I1206 11:14:15.273811 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:14:15 crc kubenswrapper[4678]: I1206 11:14:15.402523 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:14:15 crc kubenswrapper[4678]: I1206 11:14:15.456802 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kjtvm"] Dec 06 11:14:15 crc kubenswrapper[4678]: I1206 11:14:15.460266 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:14:16 crc kubenswrapper[4678]: I1206 11:14:16.404249 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kjtvm" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="registry-server" containerID="cri-o://8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b" gracePeriod=2 Dec 06 11:14:16 crc kubenswrapper[4678]: I1206 11:14:16.889708 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.034993 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-utilities\") pod \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.035173 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-catalog-content\") pod \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.035451 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tgn6\" (UniqueName: \"kubernetes.io/projected/8e30186e-a4f8-4480-90e3-cfa66d3b200b-kube-api-access-2tgn6\") pod \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\" (UID: \"8e30186e-a4f8-4480-90e3-cfa66d3b200b\") " Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.036141 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-utilities" (OuterVolumeSpecName: "utilities") pod "8e30186e-a4f8-4480-90e3-cfa66d3b200b" (UID: "8e30186e-a4f8-4480-90e3-cfa66d3b200b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.038355 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.049884 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e30186e-a4f8-4480-90e3-cfa66d3b200b-kube-api-access-2tgn6" (OuterVolumeSpecName: "kube-api-access-2tgn6") pod "8e30186e-a4f8-4480-90e3-cfa66d3b200b" (UID: "8e30186e-a4f8-4480-90e3-cfa66d3b200b"). InnerVolumeSpecName "kube-api-access-2tgn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.104298 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e30186e-a4f8-4480-90e3-cfa66d3b200b" (UID: "8e30186e-a4f8-4480-90e3-cfa66d3b200b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.141549 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e30186e-a4f8-4480-90e3-cfa66d3b200b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.141590 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tgn6\" (UniqueName: \"kubernetes.io/projected/8e30186e-a4f8-4480-90e3-cfa66d3b200b-kube-api-access-2tgn6\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.413500 4678 generic.go:334] "Generic (PLEG): container finished" podID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerID="8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b" exitCode=0 Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.413539 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjtvm" event={"ID":"8e30186e-a4f8-4480-90e3-cfa66d3b200b","Type":"ContainerDied","Data":"8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b"} Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.413587 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kjtvm" event={"ID":"8e30186e-a4f8-4480-90e3-cfa66d3b200b","Type":"ContainerDied","Data":"81d8ddfd27d1357a6eca4da12c17db26fa89f6ae99df09df278a3a62b0e6696c"} Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.413607 4678 scope.go:117] "RemoveContainer" containerID="8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.414539 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kjtvm" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.434564 4678 scope.go:117] "RemoveContainer" containerID="fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.455136 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kjtvm"] Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.465547 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kjtvm"] Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.472444 4678 scope.go:117] "RemoveContainer" containerID="203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.497368 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" path="/var/lib/kubelet/pods/8e30186e-a4f8-4480-90e3-cfa66d3b200b/volumes" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.516958 4678 scope.go:117] "RemoveContainer" containerID="8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b" Dec 06 11:14:17 crc kubenswrapper[4678]: E1206 11:14:17.517599 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b\": container with ID starting with 8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b not found: ID does not exist" containerID="8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.517649 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b"} err="failed to get container status \"8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b\": rpc error: code = NotFound desc = could not find container \"8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b\": container with ID starting with 8a610715a99931f76f05215cf89af5e4a526e2e28d0c9893204ce3ecb145996b not found: ID does not exist" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.517676 4678 scope.go:117] "RemoveContainer" containerID="fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa" Dec 06 11:14:17 crc kubenswrapper[4678]: E1206 11:14:17.518031 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa\": container with ID starting with fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa not found: ID does not exist" containerID="fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.518083 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa"} err="failed to get container status \"fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa\": rpc error: code = NotFound desc = could not find container \"fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa\": container with ID starting with fe126cab8dcd2738007708beeed4651ce56f5302db2a268a7c15a091f30552fa not found: ID does not exist" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.518112 4678 scope.go:117] "RemoveContainer" containerID="203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631" Dec 06 11:14:17 crc kubenswrapper[4678]: E1206 11:14:17.518401 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631\": container with ID starting with 203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631 not found: ID does not exist" containerID="203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.518429 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631"} err="failed to get container status \"203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631\": rpc error: code = NotFound desc = could not find container \"203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631\": container with ID starting with 203f8e5502341ddedb80083c193ad85f53d1f2721c54060bcc780cb897796631 not found: ID does not exist" Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.668322 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9vznd"] Dec 06 11:14:17 crc kubenswrapper[4678]: I1206 11:14:17.668953 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9vznd" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="registry-server" containerID="cri-o://3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064" gracePeriod=2 Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.072288 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.160980 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-utilities\") pod \"c976648c-0d90-42c1-b2ed-5071fa35b85d\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.161203 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-catalog-content\") pod \"c976648c-0d90-42c1-b2ed-5071fa35b85d\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.161252 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rq49\" (UniqueName: \"kubernetes.io/projected/c976648c-0d90-42c1-b2ed-5071fa35b85d-kube-api-access-7rq49\") pod \"c976648c-0d90-42c1-b2ed-5071fa35b85d\" (UID: \"c976648c-0d90-42c1-b2ed-5071fa35b85d\") " Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.161643 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-utilities" (OuterVolumeSpecName: "utilities") pod "c976648c-0d90-42c1-b2ed-5071fa35b85d" (UID: "c976648c-0d90-42c1-b2ed-5071fa35b85d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.176465 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c976648c-0d90-42c1-b2ed-5071fa35b85d-kube-api-access-7rq49" (OuterVolumeSpecName: "kube-api-access-7rq49") pod "c976648c-0d90-42c1-b2ed-5071fa35b85d" (UID: "c976648c-0d90-42c1-b2ed-5071fa35b85d"). InnerVolumeSpecName "kube-api-access-7rq49". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.258662 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c976648c-0d90-42c1-b2ed-5071fa35b85d" (UID: "c976648c-0d90-42c1-b2ed-5071fa35b85d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.263125 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.263160 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rq49\" (UniqueName: \"kubernetes.io/projected/c976648c-0d90-42c1-b2ed-5071fa35b85d-kube-api-access-7rq49\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.263180 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c976648c-0d90-42c1-b2ed-5071fa35b85d-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.423830 4678 generic.go:334] "Generic (PLEG): container finished" podID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerID="3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064" exitCode=0 Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.423959 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9vznd" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.428597 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vznd" event={"ID":"c976648c-0d90-42c1-b2ed-5071fa35b85d","Type":"ContainerDied","Data":"3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064"} Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.428669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9vznd" event={"ID":"c976648c-0d90-42c1-b2ed-5071fa35b85d","Type":"ContainerDied","Data":"fab178dabb0e9f29c904bba0aa80f8bb389afdf52ed9f21da6aba041edd36d1f"} Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.428688 4678 scope.go:117] "RemoveContainer" containerID="3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.468668 4678 scope.go:117] "RemoveContainer" containerID="e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.470834 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9vznd"] Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.478553 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9vznd"] Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.499227 4678 scope.go:117] "RemoveContainer" containerID="f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.526278 4678 scope.go:117] "RemoveContainer" containerID="3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064" Dec 06 11:14:18 crc kubenswrapper[4678]: E1206 11:14:18.526867 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064\": container with ID starting with 3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064 not found: ID does not exist" containerID="3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.526955 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064"} err="failed to get container status \"3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064\": rpc error: code = NotFound desc = could not find container \"3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064\": container with ID starting with 3c6a0d35c35ade1dd27b426c0322fd190d119c0bd96ee3e5bc25caf632c63064 not found: ID does not exist" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.527034 4678 scope.go:117] "RemoveContainer" containerID="e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a" Dec 06 11:14:18 crc kubenswrapper[4678]: E1206 11:14:18.529154 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a\": container with ID starting with e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a not found: ID does not exist" containerID="e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.529269 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a"} err="failed to get container status \"e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a\": rpc error: code = NotFound desc = could not find container \"e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a\": container with ID starting with e0836396f0c134fd91fa02017475eb88cd2d59142d13758ae0fdea45e9e1f69a not found: ID does not exist" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.529532 4678 scope.go:117] "RemoveContainer" containerID="f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62" Dec 06 11:14:18 crc kubenswrapper[4678]: E1206 11:14:18.529934 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62\": container with ID starting with f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62 not found: ID does not exist" containerID="f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62" Dec 06 11:14:18 crc kubenswrapper[4678]: I1206 11:14:18.529976 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62"} err="failed to get container status \"f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62\": rpc error: code = NotFound desc = could not find container \"f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62\": container with ID starting with f262ba03a78d6e8e89008bdd13ee482c9fdf6b8cc735fe724890fd171fd14b62 not found: ID does not exist" Dec 06 11:14:19 crc kubenswrapper[4678]: I1206 11:14:19.488276 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" path="/var/lib/kubelet/pods/c976648c-0d90-42c1-b2ed-5071fa35b85d/volumes" Dec 06 11:14:29 crc kubenswrapper[4678]: I1206 11:14:29.505907 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:14:29 crc kubenswrapper[4678]: I1206 11:14:29.506334 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:14:29 crc kubenswrapper[4678]: I1206 11:14:29.506375 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:14:29 crc kubenswrapper[4678]: I1206 11:14:29.507053 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:14:29 crc kubenswrapper[4678]: I1206 11:14:29.507099 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" gracePeriod=600 Dec 06 11:14:29 crc kubenswrapper[4678]: E1206 11:14:29.638082 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:14:30 crc kubenswrapper[4678]: I1206 11:14:30.542632 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" exitCode=0 Dec 06 11:14:30 crc kubenswrapper[4678]: I1206 11:14:30.542694 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5"} Dec 06 11:14:30 crc kubenswrapper[4678]: I1206 11:14:30.542961 4678 scope.go:117] "RemoveContainer" containerID="8b5b29cf01b4f5ab8126b1632136a603737bbd297541220236e9787fb95d5818" Dec 06 11:14:30 crc kubenswrapper[4678]: I1206 11:14:30.543678 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:14:30 crc kubenswrapper[4678]: E1206 11:14:30.543960 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:14:45 crc kubenswrapper[4678]: I1206 11:14:45.475664 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:14:45 crc kubenswrapper[4678]: E1206 11:14:45.476441 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:14:57 crc kubenswrapper[4678]: I1206 11:14:57.476845 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:14:57 crc kubenswrapper[4678]: E1206 11:14:57.477781 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.160656 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7"] Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.162754 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="extract-utilities" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.162860 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="extract-utilities" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.162938 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="extract-content" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.163008 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="extract-content" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.163091 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="extract-content" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.163161 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="extract-content" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.163236 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="extract-content" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.163307 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="extract-content" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.163379 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.163449 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.163594 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.163672 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.163753 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.163820 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.163898 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="extract-utilities" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.164007 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="extract-utilities" Dec 06 11:15:00 crc kubenswrapper[4678]: E1206 11:15:00.164084 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="extract-utilities" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.164152 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="extract-utilities" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.164425 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e30186e-a4f8-4480-90e3-cfa66d3b200b" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.164533 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b7f1288-f94d-4777-82dd-7e45f38fc2f3" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.164623 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c976648c-0d90-42c1-b2ed-5071fa35b85d" containerName="registry-server" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.165470 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.168348 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.171611 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.214754 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7"] Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.307417 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-secret-volume\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.308521 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-config-volume\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.308643 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsdvr\" (UniqueName: \"kubernetes.io/projected/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-kube-api-access-lsdvr\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.410125 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-config-volume\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.410205 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsdvr\" (UniqueName: \"kubernetes.io/projected/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-kube-api-access-lsdvr\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.410277 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-secret-volume\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.413264 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-config-volume\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.416457 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-secret-volume\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.428251 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsdvr\" (UniqueName: \"kubernetes.io/projected/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-kube-api-access-lsdvr\") pod \"collect-profiles-29416995-c9nt7\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:00 crc kubenswrapper[4678]: I1206 11:15:00.524021 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:01 crc kubenswrapper[4678]: I1206 11:15:01.001581 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7"] Dec 06 11:15:01 crc kubenswrapper[4678]: I1206 11:15:01.874381 4678 generic.go:334] "Generic (PLEG): container finished" podID="182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" containerID="00c4679b1e037722d88f9c27bc56eefd1a6863fca497487c2c7f860345bc243f" exitCode=0 Dec 06 11:15:01 crc kubenswrapper[4678]: I1206 11:15:01.874600 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" event={"ID":"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa","Type":"ContainerDied","Data":"00c4679b1e037722d88f9c27bc56eefd1a6863fca497487c2c7f860345bc243f"} Dec 06 11:15:01 crc kubenswrapper[4678]: I1206 11:15:01.874874 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" event={"ID":"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa","Type":"ContainerStarted","Data":"0b4c058b524814e44af667a6e21260bd8cfa6c58045d5f56e2e848f8d02a8158"} Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.283070 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.392414 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-config-volume\") pod \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.392476 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsdvr\" (UniqueName: \"kubernetes.io/projected/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-kube-api-access-lsdvr\") pod \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.392607 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-secret-volume\") pod \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\" (UID: \"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa\") " Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.393341 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-config-volume" (OuterVolumeSpecName: "config-volume") pod "182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" (UID: "182ea1b6-9b81-4b6b-ac80-7f6e0be788fa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.399212 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-kube-api-access-lsdvr" (OuterVolumeSpecName: "kube-api-access-lsdvr") pod "182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" (UID: "182ea1b6-9b81-4b6b-ac80-7f6e0be788fa"). InnerVolumeSpecName "kube-api-access-lsdvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.401641 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" (UID: "182ea1b6-9b81-4b6b-ac80-7f6e0be788fa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.494586 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.494641 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsdvr\" (UniqueName: \"kubernetes.io/projected/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-kube-api-access-lsdvr\") on node \"crc\" DevicePath \"\"" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.494664 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.902025 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" event={"ID":"182ea1b6-9b81-4b6b-ac80-7f6e0be788fa","Type":"ContainerDied","Data":"0b4c058b524814e44af667a6e21260bd8cfa6c58045d5f56e2e848f8d02a8158"} Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.902360 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b4c058b524814e44af667a6e21260bd8cfa6c58045d5f56e2e848f8d02a8158" Dec 06 11:15:03 crc kubenswrapper[4678]: I1206 11:15:03.902610 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7" Dec 06 11:15:04 crc kubenswrapper[4678]: I1206 11:15:04.353873 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6"] Dec 06 11:15:04 crc kubenswrapper[4678]: I1206 11:15:04.361074 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416950-j8mc6"] Dec 06 11:15:05 crc kubenswrapper[4678]: I1206 11:15:05.489036 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4777f215-78ef-47d4-be38-0e2a750168fe" path="/var/lib/kubelet/pods/4777f215-78ef-47d4-be38-0e2a750168fe/volumes" Dec 06 11:15:10 crc kubenswrapper[4678]: I1206 11:15:10.476007 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:15:10 crc kubenswrapper[4678]: E1206 11:15:10.478566 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:15:22 crc kubenswrapper[4678]: I1206 11:15:22.476234 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:15:22 crc kubenswrapper[4678]: E1206 11:15:22.476894 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:15:35 crc kubenswrapper[4678]: I1206 11:15:35.489160 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:15:35 crc kubenswrapper[4678]: E1206 11:15:35.491225 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:15:47 crc kubenswrapper[4678]: I1206 11:15:47.810660 4678 scope.go:117] "RemoveContainer" containerID="1ad63fcf3ea70355e95429e928b0d8ad6bb8651a83bdf2cf72bf9cc82a179b61" Dec 06 11:15:49 crc kubenswrapper[4678]: I1206 11:15:49.476089 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:15:49 crc kubenswrapper[4678]: E1206 11:15:49.477387 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:16:02 crc kubenswrapper[4678]: I1206 11:16:02.476081 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:16:02 crc kubenswrapper[4678]: E1206 11:16:02.476872 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:16:09 crc kubenswrapper[4678]: I1206 11:16:09.528987 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-f5f494679-7bm9s" podUID="b31f2128-db6e-498b-bd38-d0b1c41a4603" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 06 11:16:14 crc kubenswrapper[4678]: I1206 11:16:14.476011 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:16:14 crc kubenswrapper[4678]: E1206 11:16:14.476872 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:16:25 crc kubenswrapper[4678]: I1206 11:16:25.481526 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:16:25 crc kubenswrapper[4678]: E1206 11:16:25.482259 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:16:39 crc kubenswrapper[4678]: I1206 11:16:39.476289 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:16:39 crc kubenswrapper[4678]: E1206 11:16:39.477006 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:16:50 crc kubenswrapper[4678]: I1206 11:16:50.476031 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:16:50 crc kubenswrapper[4678]: E1206 11:16:50.476783 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:17:04 crc kubenswrapper[4678]: I1206 11:17:04.476972 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:17:04 crc kubenswrapper[4678]: E1206 11:17:04.478170 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:17:17 crc kubenswrapper[4678]: I1206 11:17:17.488049 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:17:17 crc kubenswrapper[4678]: E1206 11:17:17.490027 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:17:29 crc kubenswrapper[4678]: I1206 11:17:29.475946 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:17:29 crc kubenswrapper[4678]: E1206 11:17:29.477507 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:17:43 crc kubenswrapper[4678]: I1206 11:17:43.476755 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:17:43 crc kubenswrapper[4678]: E1206 11:17:43.477688 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:17:58 crc kubenswrapper[4678]: I1206 11:17:58.476333 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:17:58 crc kubenswrapper[4678]: E1206 11:17:58.477008 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:18:13 crc kubenswrapper[4678]: I1206 11:18:13.476410 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:18:13 crc kubenswrapper[4678]: E1206 11:18:13.477397 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:18:27 crc kubenswrapper[4678]: I1206 11:18:27.477285 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:18:27 crc kubenswrapper[4678]: E1206 11:18:27.478078 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:18:41 crc kubenswrapper[4678]: I1206 11:18:41.118072 4678 generic.go:334] "Generic (PLEG): container finished" podID="ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" containerID="c37d25a0f03941ed056beedfc2d37cf8f7571bb135eec3b7656777d7c5e5b3f3" exitCode=0 Dec 06 11:18:41 crc kubenswrapper[4678]: I1206 11:18:41.118177 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" event={"ID":"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8","Type":"ContainerDied","Data":"c37d25a0f03941ed056beedfc2d37cf8f7571bb135eec3b7656777d7c5e5b3f3"} Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.475777 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:18:42 crc kubenswrapper[4678]: E1206 11:18:42.476425 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.514914 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.637871 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-inventory\") pod \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.637910 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-ssh-key\") pod \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.637995 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-secret-0\") pod \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.638102 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-combined-ca-bundle\") pod \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.638148 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6ggl\" (UniqueName: \"kubernetes.io/projected/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-kube-api-access-l6ggl\") pod \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\" (UID: \"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8\") " Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.642848 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" (UID: "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.650715 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-kube-api-access-l6ggl" (OuterVolumeSpecName: "kube-api-access-l6ggl") pod "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" (UID: "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8"). InnerVolumeSpecName "kube-api-access-l6ggl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.666610 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" (UID: "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.669731 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-inventory" (OuterVolumeSpecName: "inventory") pod "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" (UID: "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.670583 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" (UID: "ff8f7b93-13e6-49a3-937a-8247a3eaf5c8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.741122 4678 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.741151 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6ggl\" (UniqueName: \"kubernetes.io/projected/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-kube-api-access-l6ggl\") on node \"crc\" DevicePath \"\"" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.741161 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.741169 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:18:42 crc kubenswrapper[4678]: I1206 11:18:42.741178 4678 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ff8f7b93-13e6-49a3-937a-8247a3eaf5c8-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.145477 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" event={"ID":"ff8f7b93-13e6-49a3-937a-8247a3eaf5c8","Type":"ContainerDied","Data":"15bf7a0fa6013d68fdbed79482db2841ded77970f28c246809f66957b5c910d0"} Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.145571 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="15bf7a0fa6013d68fdbed79482db2841ded77970f28c246809f66957b5c910d0" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.145973 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.332009 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld"] Dec 06 11:18:43 crc kubenswrapper[4678]: E1206 11:18:43.332469 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.332508 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 11:18:43 crc kubenswrapper[4678]: E1206 11:18:43.332534 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" containerName="collect-profiles" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.332544 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" containerName="collect-profiles" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.332793 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" containerName="collect-profiles" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.332816 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff8f7b93-13e6-49a3-937a-8247a3eaf5c8" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.333571 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.342390 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.342426 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.342448 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.342734 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.344082 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.344744 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.345810 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.351239 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld"] Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454142 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454194 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454254 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x725g\" (UniqueName: \"kubernetes.io/projected/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-kube-api-access-x725g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454295 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454382 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454433 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454605 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454707 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.454765 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.555919 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556017 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556059 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556116 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556162 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556237 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556272 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556312 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x725g\" (UniqueName: \"kubernetes.io/projected/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-kube-api-access-x725g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.556371 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.558327 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.562645 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.563136 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.564736 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.566801 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.568582 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.570041 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.571383 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.580864 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x725g\" (UniqueName: \"kubernetes.io/projected/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-kube-api-access-x725g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-4fnld\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:43 crc kubenswrapper[4678]: I1206 11:18:43.657178 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:18:44 crc kubenswrapper[4678]: I1206 11:18:44.260437 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld"] Dec 06 11:18:44 crc kubenswrapper[4678]: I1206 11:18:44.281108 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:18:45 crc kubenswrapper[4678]: I1206 11:18:45.173777 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" event={"ID":"ff53c201-513d-4b60-b0ef-a1a59a30bd4f","Type":"ContainerStarted","Data":"54d8ff776ec5daefb95ac4cd322705d1f2c180e586a175fad7afd1b02e15008f"} Dec 06 11:18:45 crc kubenswrapper[4678]: I1206 11:18:45.174055 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" event={"ID":"ff53c201-513d-4b60-b0ef-a1a59a30bd4f","Type":"ContainerStarted","Data":"e4a1000cc733992a881155ce5d1fa6153b657985b82f0357b228aeebea517c04"} Dec 06 11:18:45 crc kubenswrapper[4678]: I1206 11:18:45.198860 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" podStartSLOduration=1.727539203 podStartE2EDuration="2.198843563s" podCreationTimestamp="2025-12-06 11:18:43 +0000 UTC" firstStartedPulling="2025-12-06 11:18:44.276956917 +0000 UTC m=+2529.120388366" lastFinishedPulling="2025-12-06 11:18:44.748261287 +0000 UTC m=+2529.591692726" observedRunningTime="2025-12-06 11:18:45.190806559 +0000 UTC m=+2530.034238008" watchObservedRunningTime="2025-12-06 11:18:45.198843563 +0000 UTC m=+2530.042275002" Dec 06 11:18:55 crc kubenswrapper[4678]: I1206 11:18:55.490399 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:18:55 crc kubenswrapper[4678]: E1206 11:18:55.491765 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:19:10 crc kubenswrapper[4678]: I1206 11:19:10.475514 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:19:10 crc kubenswrapper[4678]: E1206 11:19:10.477230 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:19:24 crc kubenswrapper[4678]: I1206 11:19:24.477882 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:19:24 crc kubenswrapper[4678]: E1206 11:19:24.478702 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:19:35 crc kubenswrapper[4678]: I1206 11:19:35.483099 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:19:36 crc kubenswrapper[4678]: I1206 11:19:36.722851 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"627949b9b1b75d4fb0e9879d68f9582b2ac50cf3cb8a553036937b3bc096d583"} Dec 06 11:21:52 crc kubenswrapper[4678]: I1206 11:21:52.955264 4678 generic.go:334] "Generic (PLEG): container finished" podID="ff53c201-513d-4b60-b0ef-a1a59a30bd4f" containerID="54d8ff776ec5daefb95ac4cd322705d1f2c180e586a175fad7afd1b02e15008f" exitCode=0 Dec 06 11:21:52 crc kubenswrapper[4678]: I1206 11:21:52.956651 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" event={"ID":"ff53c201-513d-4b60-b0ef-a1a59a30bd4f","Type":"ContainerDied","Data":"54d8ff776ec5daefb95ac4cd322705d1f2c180e586a175fad7afd1b02e15008f"} Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.432342 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482366 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x725g\" (UniqueName: \"kubernetes.io/projected/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-kube-api-access-x725g\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482510 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-0\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482547 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-extra-config-0\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482610 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-ssh-key\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482672 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-0\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482688 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-inventory\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482717 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-1\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482735 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-1\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.482770 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-combined-ca-bundle\") pod \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\" (UID: \"ff53c201-513d-4b60-b0ef-a1a59a30bd4f\") " Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.512568 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-kube-api-access-x725g" (OuterVolumeSpecName: "kube-api-access-x725g") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "kube-api-access-x725g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.536070 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.556752 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.565122 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-inventory" (OuterVolumeSpecName: "inventory") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.569101 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.571173 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.572525 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.584575 4678 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.584602 4678 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.584611 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.584621 4678 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.584630 4678 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.584641 4678 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.584650 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x725g\" (UniqueName: \"kubernetes.io/projected/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-kube-api-access-x725g\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.593617 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.597250 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "ff53c201-513d-4b60-b0ef-a1a59a30bd4f" (UID: "ff53c201-513d-4b60-b0ef-a1a59a30bd4f"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.686764 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:54 crc kubenswrapper[4678]: I1206 11:21:54.686812 4678 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/ff53c201-513d-4b60-b0ef-a1a59a30bd4f-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.022708 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" event={"ID":"ff53c201-513d-4b60-b0ef-a1a59a30bd4f","Type":"ContainerDied","Data":"e4a1000cc733992a881155ce5d1fa6153b657985b82f0357b228aeebea517c04"} Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.022764 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4a1000cc733992a881155ce5d1fa6153b657985b82f0357b228aeebea517c04" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.022843 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-4fnld" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.076885 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678"] Dec 06 11:21:55 crc kubenswrapper[4678]: E1206 11:21:55.078206 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff53c201-513d-4b60-b0ef-a1a59a30bd4f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.078233 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff53c201-513d-4b60-b0ef-a1a59a30bd4f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.078813 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff53c201-513d-4b60-b0ef-a1a59a30bd4f" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.079649 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.085798 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.085909 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-mbsms" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.085977 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.086086 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.086191 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.088568 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678"] Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.198902 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.198998 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.199051 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtltw\" (UniqueName: \"kubernetes.io/projected/1cadf619-4b68-44a3-a65b-422bea0a2ede-kube-api-access-vtltw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.199104 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.199126 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.199172 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.199220 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.300432 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.300746 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.300775 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.300801 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.300872 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtltw\" (UniqueName: \"kubernetes.io/projected/1cadf619-4b68-44a3-a65b-422bea0a2ede-kube-api-access-vtltw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.300917 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.300935 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.304189 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.305590 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.306201 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.308123 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.308265 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.310661 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.319955 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtltw\" (UniqueName: \"kubernetes.io/projected/1cadf619-4b68-44a3-a65b-422bea0a2ede-kube-api-access-vtltw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-dj678\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.399870 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:21:55 crc kubenswrapper[4678]: W1206 11:21:55.973413 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1cadf619_4b68_44a3_a65b_422bea0a2ede.slice/crio-901602744a26af34d6df2ed5a8a4d2f5354e2f56e78c6145e339b69019f434dc WatchSource:0}: Error finding container 901602744a26af34d6df2ed5a8a4d2f5354e2f56e78c6145e339b69019f434dc: Status 404 returned error can't find the container with id 901602744a26af34d6df2ed5a8a4d2f5354e2f56e78c6145e339b69019f434dc Dec 06 11:21:55 crc kubenswrapper[4678]: I1206 11:21:55.991858 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678"] Dec 06 11:21:56 crc kubenswrapper[4678]: I1206 11:21:56.032581 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" event={"ID":"1cadf619-4b68-44a3-a65b-422bea0a2ede","Type":"ContainerStarted","Data":"901602744a26af34d6df2ed5a8a4d2f5354e2f56e78c6145e339b69019f434dc"} Dec 06 11:21:58 crc kubenswrapper[4678]: I1206 11:21:58.049949 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" event={"ID":"1cadf619-4b68-44a3-a65b-422bea0a2ede","Type":"ContainerStarted","Data":"6355eb0fe64cdfacadb469dd707f63637a58a000485ad8c81d47fdaf68288dc0"} Dec 06 11:21:58 crc kubenswrapper[4678]: I1206 11:21:58.079823 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" podStartSLOduration=2.370028042 podStartE2EDuration="3.07980412s" podCreationTimestamp="2025-12-06 11:21:55 +0000 UTC" firstStartedPulling="2025-12-06 11:21:55.976697611 +0000 UTC m=+2720.820129050" lastFinishedPulling="2025-12-06 11:21:56.686473689 +0000 UTC m=+2721.529905128" observedRunningTime="2025-12-06 11:21:58.076813526 +0000 UTC m=+2722.920244975" watchObservedRunningTime="2025-12-06 11:21:58.07980412 +0000 UTC m=+2722.923235579" Dec 06 11:21:59 crc kubenswrapper[4678]: I1206 11:21:59.506779 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:21:59 crc kubenswrapper[4678]: I1206 11:21:59.507108 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:22:29 crc kubenswrapper[4678]: I1206 11:22:29.506079 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:22:29 crc kubenswrapper[4678]: I1206 11:22:29.506994 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:22:59 crc kubenswrapper[4678]: I1206 11:22:59.505782 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:22:59 crc kubenswrapper[4678]: I1206 11:22:59.506525 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:22:59 crc kubenswrapper[4678]: I1206 11:22:59.506631 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:22:59 crc kubenswrapper[4678]: I1206 11:22:59.507563 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"627949b9b1b75d4fb0e9879d68f9582b2ac50cf3cb8a553036937b3bc096d583"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:22:59 crc kubenswrapper[4678]: I1206 11:22:59.507643 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://627949b9b1b75d4fb0e9879d68f9582b2ac50cf3cb8a553036937b3bc096d583" gracePeriod=600 Dec 06 11:23:00 crc kubenswrapper[4678]: I1206 11:23:00.647565 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="627949b9b1b75d4fb0e9879d68f9582b2ac50cf3cb8a553036937b3bc096d583" exitCode=0 Dec 06 11:23:00 crc kubenswrapper[4678]: I1206 11:23:00.648146 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"627949b9b1b75d4fb0e9879d68f9582b2ac50cf3cb8a553036937b3bc096d583"} Dec 06 11:23:00 crc kubenswrapper[4678]: I1206 11:23:00.648179 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73"} Dec 06 11:23:00 crc kubenswrapper[4678]: I1206 11:23:00.648199 4678 scope.go:117] "RemoveContainer" containerID="d591a6520005aa1d614468b81515589e1aa711e01403a2b4f6448d61c0fefbb5" Dec 06 11:24:01 crc kubenswrapper[4678]: I1206 11:24:01.837506 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7g5sl"] Dec 06 11:24:01 crc kubenswrapper[4678]: I1206 11:24:01.843506 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:01 crc kubenswrapper[4678]: I1206 11:24:01.935375 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7g5sl"] Dec 06 11:24:01 crc kubenswrapper[4678]: I1206 11:24:01.996502 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krdj7\" (UniqueName: \"kubernetes.io/projected/c9ad7b86-543e-4104-a371-3c2ae7a44a60-kube-api-access-krdj7\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:01 crc kubenswrapper[4678]: I1206 11:24:01.996819 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-catalog-content\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:01 crc kubenswrapper[4678]: I1206 11:24:01.996938 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-utilities\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.099081 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krdj7\" (UniqueName: \"kubernetes.io/projected/c9ad7b86-543e-4104-a371-3c2ae7a44a60-kube-api-access-krdj7\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.099397 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-catalog-content\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.099508 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-utilities\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.100233 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-utilities\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.100671 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-catalog-content\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.125409 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krdj7\" (UniqueName: \"kubernetes.io/projected/c9ad7b86-543e-4104-a371-3c2ae7a44a60-kube-api-access-krdj7\") pod \"certified-operators-7g5sl\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.188614 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:02 crc kubenswrapper[4678]: I1206 11:24:02.788101 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7g5sl"] Dec 06 11:24:03 crc kubenswrapper[4678]: I1206 11:24:03.247279 4678 generic.go:334] "Generic (PLEG): container finished" podID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerID="a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b" exitCode=0 Dec 06 11:24:03 crc kubenswrapper[4678]: I1206 11:24:03.247354 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g5sl" event={"ID":"c9ad7b86-543e-4104-a371-3c2ae7a44a60","Type":"ContainerDied","Data":"a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b"} Dec 06 11:24:03 crc kubenswrapper[4678]: I1206 11:24:03.247611 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g5sl" event={"ID":"c9ad7b86-543e-4104-a371-3c2ae7a44a60","Type":"ContainerStarted","Data":"2699538d526f1a77780317dbba68b715586e368cc646b9186cfc98539b3471c5"} Dec 06 11:24:03 crc kubenswrapper[4678]: I1206 11:24:03.249829 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:24:04 crc kubenswrapper[4678]: I1206 11:24:04.257244 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g5sl" event={"ID":"c9ad7b86-543e-4104-a371-3c2ae7a44a60","Type":"ContainerStarted","Data":"0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb"} Dec 06 11:24:06 crc kubenswrapper[4678]: I1206 11:24:06.278348 4678 generic.go:334] "Generic (PLEG): container finished" podID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerID="0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb" exitCode=0 Dec 06 11:24:06 crc kubenswrapper[4678]: I1206 11:24:06.278445 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g5sl" event={"ID":"c9ad7b86-543e-4104-a371-3c2ae7a44a60","Type":"ContainerDied","Data":"0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb"} Dec 06 11:24:08 crc kubenswrapper[4678]: I1206 11:24:08.298510 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g5sl" event={"ID":"c9ad7b86-543e-4104-a371-3c2ae7a44a60","Type":"ContainerStarted","Data":"0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07"} Dec 06 11:24:08 crc kubenswrapper[4678]: I1206 11:24:08.319120 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7g5sl" podStartSLOduration=3.356408049 podStartE2EDuration="7.31910201s" podCreationTimestamp="2025-12-06 11:24:01 +0000 UTC" firstStartedPulling="2025-12-06 11:24:03.249543154 +0000 UTC m=+2848.092974603" lastFinishedPulling="2025-12-06 11:24:07.212236705 +0000 UTC m=+2852.055668564" observedRunningTime="2025-12-06 11:24:08.31528474 +0000 UTC m=+2853.158716239" watchObservedRunningTime="2025-12-06 11:24:08.31910201 +0000 UTC m=+2853.162533449" Dec 06 11:24:12 crc kubenswrapper[4678]: I1206 11:24:12.189263 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:12 crc kubenswrapper[4678]: I1206 11:24:12.190460 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:12 crc kubenswrapper[4678]: I1206 11:24:12.243084 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:12 crc kubenswrapper[4678]: I1206 11:24:12.392345 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:12 crc kubenswrapper[4678]: I1206 11:24:12.476660 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7g5sl"] Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.367176 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7g5sl" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="registry-server" containerID="cri-o://0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07" gracePeriod=2 Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.778991 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.958178 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-utilities\") pod \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.958577 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krdj7\" (UniqueName: \"kubernetes.io/projected/c9ad7b86-543e-4104-a371-3c2ae7a44a60-kube-api-access-krdj7\") pod \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.958736 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-catalog-content\") pod \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\" (UID: \"c9ad7b86-543e-4104-a371-3c2ae7a44a60\") " Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.959027 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-utilities" (OuterVolumeSpecName: "utilities") pod "c9ad7b86-543e-4104-a371-3c2ae7a44a60" (UID: "c9ad7b86-543e-4104-a371-3c2ae7a44a60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.959588 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:24:14 crc kubenswrapper[4678]: I1206 11:24:14.968623 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9ad7b86-543e-4104-a371-3c2ae7a44a60-kube-api-access-krdj7" (OuterVolumeSpecName: "kube-api-access-krdj7") pod "c9ad7b86-543e-4104-a371-3c2ae7a44a60" (UID: "c9ad7b86-543e-4104-a371-3c2ae7a44a60"). InnerVolumeSpecName "kube-api-access-krdj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.046326 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9ad7b86-543e-4104-a371-3c2ae7a44a60" (UID: "c9ad7b86-543e-4104-a371-3c2ae7a44a60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.061564 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krdj7\" (UniqueName: \"kubernetes.io/projected/c9ad7b86-543e-4104-a371-3c2ae7a44a60-kube-api-access-krdj7\") on node \"crc\" DevicePath \"\"" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.061658 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9ad7b86-543e-4104-a371-3c2ae7a44a60-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.379243 4678 generic.go:334] "Generic (PLEG): container finished" podID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerID="0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07" exitCode=0 Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.379289 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7g5sl" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.379306 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g5sl" event={"ID":"c9ad7b86-543e-4104-a371-3c2ae7a44a60","Type":"ContainerDied","Data":"0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07"} Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.379344 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7g5sl" event={"ID":"c9ad7b86-543e-4104-a371-3c2ae7a44a60","Type":"ContainerDied","Data":"2699538d526f1a77780317dbba68b715586e368cc646b9186cfc98539b3471c5"} Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.379364 4678 scope.go:117] "RemoveContainer" containerID="0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.420053 4678 scope.go:117] "RemoveContainer" containerID="0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.458701 4678 scope.go:117] "RemoveContainer" containerID="a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.458855 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7g5sl"] Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.467587 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7g5sl"] Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.490046 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" path="/var/lib/kubelet/pods/c9ad7b86-543e-4104-a371-3c2ae7a44a60/volumes" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.514717 4678 scope.go:117] "RemoveContainer" containerID="0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07" Dec 06 11:24:15 crc kubenswrapper[4678]: E1206 11:24:15.515202 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07\": container with ID starting with 0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07 not found: ID does not exist" containerID="0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.515236 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07"} err="failed to get container status \"0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07\": rpc error: code = NotFound desc = could not find container \"0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07\": container with ID starting with 0891525004e7969db4d5474e39e237840015ea2a7c01824d56b4ec664a6c1d07 not found: ID does not exist" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.515258 4678 scope.go:117] "RemoveContainer" containerID="0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb" Dec 06 11:24:15 crc kubenswrapper[4678]: E1206 11:24:15.515702 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb\": container with ID starting with 0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb not found: ID does not exist" containerID="0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.515785 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb"} err="failed to get container status \"0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb\": rpc error: code = NotFound desc = could not find container \"0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb\": container with ID starting with 0d01fbf1f8b0582d957104f28f253fcc936c7fe163870614db87f70199714cfb not found: ID does not exist" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.515814 4678 scope.go:117] "RemoveContainer" containerID="a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b" Dec 06 11:24:15 crc kubenswrapper[4678]: E1206 11:24:15.516270 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b\": container with ID starting with a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b not found: ID does not exist" containerID="a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b" Dec 06 11:24:15 crc kubenswrapper[4678]: I1206 11:24:15.516307 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b"} err="failed to get container status \"a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b\": rpc error: code = NotFound desc = could not find container \"a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b\": container with ID starting with a8cd39c54514d637c8abf088a9c1c12d6c5af6b80e58f62076c9032901c3ea1b not found: ID does not exist" Dec 06 11:24:59 crc kubenswrapper[4678]: I1206 11:24:59.505598 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:24:59 crc kubenswrapper[4678]: I1206 11:24:59.506198 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:25:29 crc kubenswrapper[4678]: I1206 11:25:29.505166 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:25:29 crc kubenswrapper[4678]: I1206 11:25:29.505638 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:25:39 crc kubenswrapper[4678]: I1206 11:25:39.302179 4678 generic.go:334] "Generic (PLEG): container finished" podID="1cadf619-4b68-44a3-a65b-422bea0a2ede" containerID="6355eb0fe64cdfacadb469dd707f63637a58a000485ad8c81d47fdaf68288dc0" exitCode=0 Dec 06 11:25:39 crc kubenswrapper[4678]: I1206 11:25:39.302255 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" event={"ID":"1cadf619-4b68-44a3-a65b-422bea0a2ede","Type":"ContainerDied","Data":"6355eb0fe64cdfacadb469dd707f63637a58a000485ad8c81d47fdaf68288dc0"} Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.790930 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.986645 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-inventory\") pod \"1cadf619-4b68-44a3-a65b-422bea0a2ede\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.986723 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-telemetry-combined-ca-bundle\") pod \"1cadf619-4b68-44a3-a65b-422bea0a2ede\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.986839 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtltw\" (UniqueName: \"kubernetes.io/projected/1cadf619-4b68-44a3-a65b-422bea0a2ede-kube-api-access-vtltw\") pod \"1cadf619-4b68-44a3-a65b-422bea0a2ede\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.986882 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-2\") pod \"1cadf619-4b68-44a3-a65b-422bea0a2ede\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.986945 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ssh-key\") pod \"1cadf619-4b68-44a3-a65b-422bea0a2ede\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.986970 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-1\") pod \"1cadf619-4b68-44a3-a65b-422bea0a2ede\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " Dec 06 11:25:40 crc kubenswrapper[4678]: I1206 11:25:40.987006 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-0\") pod \"1cadf619-4b68-44a3-a65b-422bea0a2ede\" (UID: \"1cadf619-4b68-44a3-a65b-422bea0a2ede\") " Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.026741 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cadf619-4b68-44a3-a65b-422bea0a2ede-kube-api-access-vtltw" (OuterVolumeSpecName: "kube-api-access-vtltw") pod "1cadf619-4b68-44a3-a65b-422bea0a2ede" (UID: "1cadf619-4b68-44a3-a65b-422bea0a2ede"). InnerVolumeSpecName "kube-api-access-vtltw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.038179 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "1cadf619-4b68-44a3-a65b-422bea0a2ede" (UID: "1cadf619-4b68-44a3-a65b-422bea0a2ede"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.056835 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "1cadf619-4b68-44a3-a65b-422bea0a2ede" (UID: "1cadf619-4b68-44a3-a65b-422bea0a2ede"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.070498 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-inventory" (OuterVolumeSpecName: "inventory") pod "1cadf619-4b68-44a3-a65b-422bea0a2ede" (UID: "1cadf619-4b68-44a3-a65b-422bea0a2ede"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.079895 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "1cadf619-4b68-44a3-a65b-422bea0a2ede" (UID: "1cadf619-4b68-44a3-a65b-422bea0a2ede"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.089140 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.089168 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.089178 4678 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-inventory\") on node \"crc\" DevicePath \"\"" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.089188 4678 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.089198 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtltw\" (UniqueName: \"kubernetes.io/projected/1cadf619-4b68-44a3-a65b-422bea0a2ede-kube-api-access-vtltw\") on node \"crc\" DevicePath \"\"" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.094428 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1cadf619-4b68-44a3-a65b-422bea0a2ede" (UID: "1cadf619-4b68-44a3-a65b-422bea0a2ede"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.094599 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "1cadf619-4b68-44a3-a65b-422bea0a2ede" (UID: "1cadf619-4b68-44a3-a65b-422bea0a2ede"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.190973 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.190999 4678 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/1cadf619-4b68-44a3-a65b-422bea0a2ede-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.322357 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" event={"ID":"1cadf619-4b68-44a3-a65b-422bea0a2ede","Type":"ContainerDied","Data":"901602744a26af34d6df2ed5a8a4d2f5354e2f56e78c6145e339b69019f434dc"} Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.322392 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="901602744a26af34d6df2ed5a8a4d2f5354e2f56e78c6145e339b69019f434dc" Dec 06 11:25:41 crc kubenswrapper[4678]: I1206 11:25:41.322858 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-dj678" Dec 06 11:25:59 crc kubenswrapper[4678]: I1206 11:25:59.505292 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:25:59 crc kubenswrapper[4678]: I1206 11:25:59.505802 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:25:59 crc kubenswrapper[4678]: I1206 11:25:59.507398 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:25:59 crc kubenswrapper[4678]: I1206 11:25:59.508105 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:25:59 crc kubenswrapper[4678]: I1206 11:25:59.508443 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" gracePeriod=600 Dec 06 11:25:59 crc kubenswrapper[4678]: E1206 11:25:59.629872 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:26:00 crc kubenswrapper[4678]: I1206 11:26:00.518699 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" exitCode=0 Dec 06 11:26:00 crc kubenswrapper[4678]: I1206 11:26:00.518870 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73"} Dec 06 11:26:00 crc kubenswrapper[4678]: I1206 11:26:00.519120 4678 scope.go:117] "RemoveContainer" containerID="627949b9b1b75d4fb0e9879d68f9582b2ac50cf3cb8a553036937b3bc096d583" Dec 06 11:26:00 crc kubenswrapper[4678]: I1206 11:26:00.519830 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:26:00 crc kubenswrapper[4678]: E1206 11:26:00.520272 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:26:11 crc kubenswrapper[4678]: I1206 11:26:11.476230 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:26:11 crc kubenswrapper[4678]: E1206 11:26:11.477371 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:26:22 crc kubenswrapper[4678]: I1206 11:26:22.476753 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:26:22 crc kubenswrapper[4678]: E1206 11:26:22.477396 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:26:36 crc kubenswrapper[4678]: I1206 11:26:36.475769 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:26:36 crc kubenswrapper[4678]: E1206 11:26:36.476634 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.857064 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 11:26:38 crc kubenswrapper[4678]: E1206 11:26:38.857810 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="registry-server" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.857827 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="registry-server" Dec 06 11:26:38 crc kubenswrapper[4678]: E1206 11:26:38.857839 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cadf619-4b68-44a3-a65b-422bea0a2ede" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.857846 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cadf619-4b68-44a3-a65b-422bea0a2ede" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 11:26:38 crc kubenswrapper[4678]: E1206 11:26:38.857862 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="extract-content" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.857868 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="extract-content" Dec 06 11:26:38 crc kubenswrapper[4678]: E1206 11:26:38.857887 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="extract-utilities" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.857893 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="extract-utilities" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.858078 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cadf619-4b68-44a3-a65b-422bea0a2ede" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.858090 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9ad7b86-543e-4104-a371-3c2ae7a44a60" containerName="registry-server" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.858855 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.862051 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.862052 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.862591 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-d84d5" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.863218 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.873953 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958172 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958506 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958543 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958598 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958635 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958662 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958683 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44qxb\" (UniqueName: \"kubernetes.io/projected/5b32697e-d8b7-4852-8afe-a6f19920f210-kube-api-access-44qxb\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958746 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-config-data\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:38 crc kubenswrapper[4678]: I1206 11:26:38.958770 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.060678 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.060765 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.060805 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.060840 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.060862 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44qxb\" (UniqueName: \"kubernetes.io/projected/5b32697e-d8b7-4852-8afe-a6f19920f210-kube-api-access-44qxb\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.060896 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-config-data\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.060928 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.061000 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.061051 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.061198 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.061524 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.061530 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.062482 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.062743 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-config-data\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.069403 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.070751 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.077352 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.083880 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44qxb\" (UniqueName: \"kubernetes.io/projected/5b32697e-d8b7-4852-8afe-a6f19920f210-kube-api-access-44qxb\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.091345 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.189816 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.686003 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 06 11:26:39 crc kubenswrapper[4678]: I1206 11:26:39.922258 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5b32697e-d8b7-4852-8afe-a6f19920f210","Type":"ContainerStarted","Data":"4628b8f635c1686c6e116c742744f1d081e792b40af640372ec7e9fd6947253d"} Dec 06 11:26:49 crc kubenswrapper[4678]: I1206 11:26:49.481371 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:26:49 crc kubenswrapper[4678]: E1206 11:26:49.482897 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:27:03 crc kubenswrapper[4678]: I1206 11:27:03.480372 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:27:03 crc kubenswrapper[4678]: E1206 11:27:03.481175 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:27:15 crc kubenswrapper[4678]: I1206 11:27:15.524466 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-f5f494679-7bm9s" podUID="b31f2128-db6e-498b-bd38-d0b1c41a4603" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 06 11:27:16 crc kubenswrapper[4678]: I1206 11:27:16.475776 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:27:16 crc kubenswrapper[4678]: E1206 11:27:16.476521 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:27:17 crc kubenswrapper[4678]: I1206 11:27:17.768170 4678 patch_prober.go:28] interesting pod/oauth-openshift-54f75f9d4b-dsnqg container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 06 11:27:17 crc kubenswrapper[4678]: I1206 11:27:17.768592 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-54f75f9d4b-dsnqg" podUID="0926d6b9-4e1c-4413-8e02-02a9d2417956" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.56:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 06 11:27:20 crc kubenswrapper[4678]: E1206 11:27:20.382814 4678 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 06 11:27:20 crc kubenswrapper[4678]: E1206 11:27:20.386684 4678 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-44qxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(5b32697e-d8b7-4852-8afe-a6f19920f210): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 06 11:27:20 crc kubenswrapper[4678]: E1206 11:27:20.387782 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="5b32697e-d8b7-4852-8afe-a6f19920f210" Dec 06 11:27:21 crc kubenswrapper[4678]: E1206 11:27:21.407881 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="5b32697e-d8b7-4852-8afe-a6f19920f210" Dec 06 11:27:28 crc kubenswrapper[4678]: I1206 11:27:28.476470 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:27:28 crc kubenswrapper[4678]: E1206 11:27:28.477262 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:27:34 crc kubenswrapper[4678]: I1206 11:27:34.527949 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5b32697e-d8b7-4852-8afe-a6f19920f210","Type":"ContainerStarted","Data":"4f70958e4fa5a118a4e6f302724c076b674246de73a65eb898bc894a669b3ff2"} Dec 06 11:27:44 crc kubenswrapper[4678]: I1206 11:27:44.476123 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:27:44 crc kubenswrapper[4678]: E1206 11:27:44.476798 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:27:58 crc kubenswrapper[4678]: I1206 11:27:58.475439 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:27:58 crc kubenswrapper[4678]: E1206 11:27:58.476070 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:28:12 crc kubenswrapper[4678]: I1206 11:28:12.477238 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:28:12 crc kubenswrapper[4678]: E1206 11:28:12.478078 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:28:27 crc kubenswrapper[4678]: I1206 11:28:27.475651 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:28:27 crc kubenswrapper[4678]: E1206 11:28:27.476357 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:28:41 crc kubenswrapper[4678]: I1206 11:28:41.476302 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:28:41 crc kubenswrapper[4678]: E1206 11:28:41.477123 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:28:52 crc kubenswrapper[4678]: I1206 11:28:52.476883 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:28:52 crc kubenswrapper[4678]: E1206 11:28:52.478946 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:29:04 crc kubenswrapper[4678]: I1206 11:29:04.475818 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:29:04 crc kubenswrapper[4678]: E1206 11:29:04.476460 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:29:17 crc kubenswrapper[4678]: I1206 11:29:17.477255 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:29:17 crc kubenswrapper[4678]: E1206 11:29:17.478030 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:29:28 crc kubenswrapper[4678]: I1206 11:29:28.475756 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:29:28 crc kubenswrapper[4678]: E1206 11:29:28.476409 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:29:39 crc kubenswrapper[4678]: I1206 11:29:39.475757 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:29:39 crc kubenswrapper[4678]: E1206 11:29:39.476519 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:29:51 crc kubenswrapper[4678]: I1206 11:29:51.476548 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:29:51 crc kubenswrapper[4678]: E1206 11:29:51.477350 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:29:53 crc kubenswrapper[4678]: I1206 11:29:53.979844 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=143.696219132 podStartE2EDuration="3m16.979824734s" podCreationTimestamp="2025-12-06 11:26:37 +0000 UTC" firstStartedPulling="2025-12-06 11:26:39.690864228 +0000 UTC m=+3004.534295707" lastFinishedPulling="2025-12-06 11:27:32.97446987 +0000 UTC m=+3057.817901309" observedRunningTime="2025-12-06 11:27:34.551834392 +0000 UTC m=+3059.395265841" watchObservedRunningTime="2025-12-06 11:29:53.979824734 +0000 UTC m=+3198.823256173" Dec 06 11:29:53 crc kubenswrapper[4678]: I1206 11:29:53.985027 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7sznb"] Dec 06 11:29:53 crc kubenswrapper[4678]: I1206 11:29:53.986864 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.043802 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sznb"] Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.056217 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-utilities\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.056326 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc9nl\" (UniqueName: \"kubernetes.io/projected/e407ea61-547c-43f5-842e-dad0c4597c07-kube-api-access-wc9nl\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.056364 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-catalog-content\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.158532 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-utilities\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.158679 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc9nl\" (UniqueName: \"kubernetes.io/projected/e407ea61-547c-43f5-842e-dad0c4597c07-kube-api-access-wc9nl\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.158723 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-catalog-content\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.159040 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-utilities\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.159390 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-catalog-content\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.190159 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc9nl\" (UniqueName: \"kubernetes.io/projected/e407ea61-547c-43f5-842e-dad0c4597c07-kube-api-access-wc9nl\") pod \"redhat-marketplace-7sznb\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:54 crc kubenswrapper[4678]: I1206 11:29:54.307973 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:29:55 crc kubenswrapper[4678]: I1206 11:29:55.047265 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sznb"] Dec 06 11:29:55 crc kubenswrapper[4678]: W1206 11:29:55.050953 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode407ea61_547c_43f5_842e_dad0c4597c07.slice/crio-09ff5aa3982dfe1f1082ec3daff093205f93699b3f4ea7121ba34a7f647ce869 WatchSource:0}: Error finding container 09ff5aa3982dfe1f1082ec3daff093205f93699b3f4ea7121ba34a7f647ce869: Status 404 returned error can't find the container with id 09ff5aa3982dfe1f1082ec3daff093205f93699b3f4ea7121ba34a7f647ce869 Dec 06 11:29:55 crc kubenswrapper[4678]: I1206 11:29:55.802783 4678 generic.go:334] "Generic (PLEG): container finished" podID="e407ea61-547c-43f5-842e-dad0c4597c07" containerID="abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f" exitCode=0 Dec 06 11:29:55 crc kubenswrapper[4678]: I1206 11:29:55.803084 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sznb" event={"ID":"e407ea61-547c-43f5-842e-dad0c4597c07","Type":"ContainerDied","Data":"abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f"} Dec 06 11:29:55 crc kubenswrapper[4678]: I1206 11:29:55.803115 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sznb" event={"ID":"e407ea61-547c-43f5-842e-dad0c4597c07","Type":"ContainerStarted","Data":"09ff5aa3982dfe1f1082ec3daff093205f93699b3f4ea7121ba34a7f647ce869"} Dec 06 11:29:55 crc kubenswrapper[4678]: I1206 11:29:55.805065 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.369789 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dnkx5"] Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.372149 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.398774 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dnkx5"] Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.504327 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-catalog-content\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.504385 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rb7pz\" (UniqueName: \"kubernetes.io/projected/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-kube-api-access-rb7pz\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.504464 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-utilities\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.572088 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-68z4l"] Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.574288 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.591852 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68z4l"] Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.606076 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-catalog-content\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.606132 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rb7pz\" (UniqueName: \"kubernetes.io/projected/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-kube-api-access-rb7pz\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.606197 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-utilities\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.607975 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-catalog-content\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.609187 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-utilities\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.634514 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rb7pz\" (UniqueName: \"kubernetes.io/projected/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-kube-api-access-rb7pz\") pod \"community-operators-dnkx5\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.692600 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.711525 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfn8h\" (UniqueName: \"kubernetes.io/projected/4aa99324-976f-4fbe-9c05-edf1f0ff783a-kube-api-access-kfn8h\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.711867 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-catalog-content\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.712001 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-utilities\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.813654 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfn8h\" (UniqueName: \"kubernetes.io/projected/4aa99324-976f-4fbe-9c05-edf1f0ff783a-kube-api-access-kfn8h\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.813727 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-catalog-content\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.813787 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-utilities\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.814356 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-utilities\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.814583 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-catalog-content\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.821320 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sznb" event={"ID":"e407ea61-547c-43f5-842e-dad0c4597c07","Type":"ContainerStarted","Data":"c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03"} Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.834829 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfn8h\" (UniqueName: \"kubernetes.io/projected/4aa99324-976f-4fbe-9c05-edf1f0ff783a-kube-api-access-kfn8h\") pod \"redhat-operators-68z4l\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:56 crc kubenswrapper[4678]: I1206 11:29:56.900958 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.321805 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dnkx5"] Dec 06 11:29:57 crc kubenswrapper[4678]: W1206 11:29:57.326111 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd17a6604_cafe_4e6d_9df5_cb6ebc12720b.slice/crio-71ed2fe114bf9a28e50078a503cbc9cef34fc0b54dbe7870a0a8a73750b74ae9 WatchSource:0}: Error finding container 71ed2fe114bf9a28e50078a503cbc9cef34fc0b54dbe7870a0a8a73750b74ae9: Status 404 returned error can't find the container with id 71ed2fe114bf9a28e50078a503cbc9cef34fc0b54dbe7870a0a8a73750b74ae9 Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.511813 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-68z4l"] Dec 06 11:29:57 crc kubenswrapper[4678]: W1206 11:29:57.525527 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aa99324_976f_4fbe_9c05_edf1f0ff783a.slice/crio-f819730d2fa67f6a6aa1832c0ef0faa7e2643ea02d9bf23bfbf206b7d180bad0 WatchSource:0}: Error finding container f819730d2fa67f6a6aa1832c0ef0faa7e2643ea02d9bf23bfbf206b7d180bad0: Status 404 returned error can't find the container with id f819730d2fa67f6a6aa1832c0ef0faa7e2643ea02d9bf23bfbf206b7d180bad0 Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.831046 4678 generic.go:334] "Generic (PLEG): container finished" podID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerID="238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd" exitCode=0 Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.831233 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68z4l" event={"ID":"4aa99324-976f-4fbe-9c05-edf1f0ff783a","Type":"ContainerDied","Data":"238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd"} Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.831344 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68z4l" event={"ID":"4aa99324-976f-4fbe-9c05-edf1f0ff783a","Type":"ContainerStarted","Data":"f819730d2fa67f6a6aa1832c0ef0faa7e2643ea02d9bf23bfbf206b7d180bad0"} Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.836119 4678 generic.go:334] "Generic (PLEG): container finished" podID="e407ea61-547c-43f5-842e-dad0c4597c07" containerID="c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03" exitCode=0 Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.836204 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sznb" event={"ID":"e407ea61-547c-43f5-842e-dad0c4597c07","Type":"ContainerDied","Data":"c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03"} Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.839341 4678 generic.go:334] "Generic (PLEG): container finished" podID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerID="52bb33290e0bb31d0522024a3bcc2e3a1df93f070df94df0133bbcaa7221c3bd" exitCode=0 Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.839392 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnkx5" event={"ID":"d17a6604-cafe-4e6d-9df5-cb6ebc12720b","Type":"ContainerDied","Data":"52bb33290e0bb31d0522024a3bcc2e3a1df93f070df94df0133bbcaa7221c3bd"} Dec 06 11:29:57 crc kubenswrapper[4678]: I1206 11:29:57.839417 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnkx5" event={"ID":"d17a6604-cafe-4e6d-9df5-cb6ebc12720b","Type":"ContainerStarted","Data":"71ed2fe114bf9a28e50078a503cbc9cef34fc0b54dbe7870a0a8a73750b74ae9"} Dec 06 11:29:58 crc kubenswrapper[4678]: I1206 11:29:58.850788 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68z4l" event={"ID":"4aa99324-976f-4fbe-9c05-edf1f0ff783a","Type":"ContainerStarted","Data":"777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd"} Dec 06 11:29:58 crc kubenswrapper[4678]: I1206 11:29:58.852917 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sznb" event={"ID":"e407ea61-547c-43f5-842e-dad0c4597c07","Type":"ContainerStarted","Data":"4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc"} Dec 06 11:29:58 crc kubenswrapper[4678]: I1206 11:29:58.856299 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnkx5" event={"ID":"d17a6604-cafe-4e6d-9df5-cb6ebc12720b","Type":"ContainerStarted","Data":"1d221d2c6f39c10763810959a570233ffb787caf1ea18646a79fcebdd590f954"} Dec 06 11:29:58 crc kubenswrapper[4678]: I1206 11:29:58.916674 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7sznb" podStartSLOduration=3.360519899 podStartE2EDuration="5.916654305s" podCreationTimestamp="2025-12-06 11:29:53 +0000 UTC" firstStartedPulling="2025-12-06 11:29:55.804870272 +0000 UTC m=+3200.648301711" lastFinishedPulling="2025-12-06 11:29:58.361004678 +0000 UTC m=+3203.204436117" observedRunningTime="2025-12-06 11:29:58.914454957 +0000 UTC m=+3203.757886396" watchObservedRunningTime="2025-12-06 11:29:58.916654305 +0000 UTC m=+3203.760085734" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.331086 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh"] Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.342840 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.350789 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.352846 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.361244 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh"] Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.492265 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsvv6\" (UniqueName: \"kubernetes.io/projected/3b175a69-f992-4f15-8257-a9a91b1b364f-kube-api-access-dsvv6\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.492321 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b175a69-f992-4f15-8257-a9a91b1b364f-config-volume\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.492383 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b175a69-f992-4f15-8257-a9a91b1b364f-secret-volume\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.593567 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b175a69-f992-4f15-8257-a9a91b1b364f-secret-volume\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.593706 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsvv6\" (UniqueName: \"kubernetes.io/projected/3b175a69-f992-4f15-8257-a9a91b1b364f-kube-api-access-dsvv6\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.593739 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b175a69-f992-4f15-8257-a9a91b1b364f-config-volume\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.594593 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b175a69-f992-4f15-8257-a9a91b1b364f-config-volume\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.618229 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsvv6\" (UniqueName: \"kubernetes.io/projected/3b175a69-f992-4f15-8257-a9a91b1b364f-kube-api-access-dsvv6\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.621710 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b175a69-f992-4f15-8257-a9a91b1b364f-secret-volume\") pod \"collect-profiles-29417010-wslfh\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:00 crc kubenswrapper[4678]: I1206 11:30:00.665091 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:01 crc kubenswrapper[4678]: I1206 11:30:01.278276 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh"] Dec 06 11:30:01 crc kubenswrapper[4678]: I1206 11:30:01.886664 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" event={"ID":"3b175a69-f992-4f15-8257-a9a91b1b364f","Type":"ContainerStarted","Data":"7ff4e054dde44b34dae6f51b70fa0cd4071d6934f0dc7676137852b1e8a7e390"} Dec 06 11:30:01 crc kubenswrapper[4678]: I1206 11:30:01.886930 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" event={"ID":"3b175a69-f992-4f15-8257-a9a91b1b364f","Type":"ContainerStarted","Data":"cbf8cf38826ebeee010b67a12ce99f6cf1f9ca845ba81604a7e154312557f3d8"} Dec 06 11:30:01 crc kubenswrapper[4678]: I1206 11:30:01.909100 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" podStartSLOduration=1.909082469 podStartE2EDuration="1.909082469s" podCreationTimestamp="2025-12-06 11:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:30:01.89982188 +0000 UTC m=+3206.743253319" watchObservedRunningTime="2025-12-06 11:30:01.909082469 +0000 UTC m=+3206.752513908" Dec 06 11:30:02 crc kubenswrapper[4678]: I1206 11:30:02.899338 4678 generic.go:334] "Generic (PLEG): container finished" podID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerID="1d221d2c6f39c10763810959a570233ffb787caf1ea18646a79fcebdd590f954" exitCode=0 Dec 06 11:30:02 crc kubenswrapper[4678]: I1206 11:30:02.899418 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnkx5" event={"ID":"d17a6604-cafe-4e6d-9df5-cb6ebc12720b","Type":"ContainerDied","Data":"1d221d2c6f39c10763810959a570233ffb787caf1ea18646a79fcebdd590f954"} Dec 06 11:30:03 crc kubenswrapper[4678]: I1206 11:30:03.476844 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:30:03 crc kubenswrapper[4678]: E1206 11:30:03.477628 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:30:03 crc kubenswrapper[4678]: I1206 11:30:03.913252 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnkx5" event={"ID":"d17a6604-cafe-4e6d-9df5-cb6ebc12720b","Type":"ContainerStarted","Data":"9e7a91291cf12f4aa21196898ca89b78353e117c5c842d660b0a692ccf55c085"} Dec 06 11:30:03 crc kubenswrapper[4678]: I1206 11:30:03.915230 4678 generic.go:334] "Generic (PLEG): container finished" podID="3b175a69-f992-4f15-8257-a9a91b1b364f" containerID="7ff4e054dde44b34dae6f51b70fa0cd4071d6934f0dc7676137852b1e8a7e390" exitCode=0 Dec 06 11:30:03 crc kubenswrapper[4678]: I1206 11:30:03.915277 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" event={"ID":"3b175a69-f992-4f15-8257-a9a91b1b364f","Type":"ContainerDied","Data":"7ff4e054dde44b34dae6f51b70fa0cd4071d6934f0dc7676137852b1e8a7e390"} Dec 06 11:30:04 crc kubenswrapper[4678]: I1206 11:30:04.308102 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:30:04 crc kubenswrapper[4678]: I1206 11:30:04.308948 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:30:04 crc kubenswrapper[4678]: I1206 11:30:04.927112 4678 generic.go:334] "Generic (PLEG): container finished" podID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerID="777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd" exitCode=0 Dec 06 11:30:04 crc kubenswrapper[4678]: I1206 11:30:04.927284 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68z4l" event={"ID":"4aa99324-976f-4fbe-9c05-edf1f0ff783a","Type":"ContainerDied","Data":"777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd"} Dec 06 11:30:04 crc kubenswrapper[4678]: I1206 11:30:04.980050 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dnkx5" podStartSLOduration=3.405729573 podStartE2EDuration="8.980027687s" podCreationTimestamp="2025-12-06 11:29:56 +0000 UTC" firstStartedPulling="2025-12-06 11:29:57.841384517 +0000 UTC m=+3202.684815956" lastFinishedPulling="2025-12-06 11:30:03.415682631 +0000 UTC m=+3208.259114070" observedRunningTime="2025-12-06 11:30:04.954706616 +0000 UTC m=+3209.798138065" watchObservedRunningTime="2025-12-06 11:30:04.980027687 +0000 UTC m=+3209.823459126" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.363870 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-7sznb" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="registry-server" probeResult="failure" output=< Dec 06 11:30:05 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:30:05 crc kubenswrapper[4678]: > Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.560563 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.618621 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsvv6\" (UniqueName: \"kubernetes.io/projected/3b175a69-f992-4f15-8257-a9a91b1b364f-kube-api-access-dsvv6\") pod \"3b175a69-f992-4f15-8257-a9a91b1b364f\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.618837 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b175a69-f992-4f15-8257-a9a91b1b364f-secret-volume\") pod \"3b175a69-f992-4f15-8257-a9a91b1b364f\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.619169 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b175a69-f992-4f15-8257-a9a91b1b364f-config-volume\") pod \"3b175a69-f992-4f15-8257-a9a91b1b364f\" (UID: \"3b175a69-f992-4f15-8257-a9a91b1b364f\") " Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.621448 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b175a69-f992-4f15-8257-a9a91b1b364f-config-volume" (OuterVolumeSpecName: "config-volume") pod "3b175a69-f992-4f15-8257-a9a91b1b364f" (UID: "3b175a69-f992-4f15-8257-a9a91b1b364f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.633394 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b175a69-f992-4f15-8257-a9a91b1b364f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3b175a69-f992-4f15-8257-a9a91b1b364f" (UID: "3b175a69-f992-4f15-8257-a9a91b1b364f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.659839 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b175a69-f992-4f15-8257-a9a91b1b364f-kube-api-access-dsvv6" (OuterVolumeSpecName: "kube-api-access-dsvv6") pod "3b175a69-f992-4f15-8257-a9a91b1b364f" (UID: "3b175a69-f992-4f15-8257-a9a91b1b364f"). InnerVolumeSpecName "kube-api-access-dsvv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.722785 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b175a69-f992-4f15-8257-a9a91b1b364f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.722822 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsvv6\" (UniqueName: \"kubernetes.io/projected/3b175a69-f992-4f15-8257-a9a91b1b364f-kube-api-access-dsvv6\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.722833 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b175a69-f992-4f15-8257-a9a91b1b364f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.937475 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68z4l" event={"ID":"4aa99324-976f-4fbe-9c05-edf1f0ff783a","Type":"ContainerStarted","Data":"c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb"} Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.939243 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" event={"ID":"3b175a69-f992-4f15-8257-a9a91b1b364f","Type":"ContainerDied","Data":"cbf8cf38826ebeee010b67a12ce99f6cf1f9ca845ba81604a7e154312557f3d8"} Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.939276 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417010-wslfh" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.939279 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbf8cf38826ebeee010b67a12ce99f6cf1f9ca845ba81604a7e154312557f3d8" Dec 06 11:30:05 crc kubenswrapper[4678]: I1206 11:30:05.978903 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-68z4l" podStartSLOduration=2.443257138 podStartE2EDuration="9.978888137s" podCreationTimestamp="2025-12-06 11:29:56 +0000 UTC" firstStartedPulling="2025-12-06 11:29:57.832962884 +0000 UTC m=+3202.676394323" lastFinishedPulling="2025-12-06 11:30:05.368593883 +0000 UTC m=+3210.212025322" observedRunningTime="2025-12-06 11:30:05.969865135 +0000 UTC m=+3210.813296574" watchObservedRunningTime="2025-12-06 11:30:05.978888137 +0000 UTC m=+3210.822319576" Dec 06 11:30:06 crc kubenswrapper[4678]: I1206 11:30:06.692948 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:30:06 crc kubenswrapper[4678]: I1206 11:30:06.694245 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:30:06 crc kubenswrapper[4678]: I1206 11:30:06.782756 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5"] Dec 06 11:30:06 crc kubenswrapper[4678]: I1206 11:30:06.790459 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416965-cj9z5"] Dec 06 11:30:06 crc kubenswrapper[4678]: I1206 11:30:06.901311 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:30:06 crc kubenswrapper[4678]: I1206 11:30:06.901496 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:30:07 crc kubenswrapper[4678]: I1206 11:30:07.487925 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33764757-2f8b-4c86-982d-4e9efde21884" path="/var/lib/kubelet/pods/33764757-2f8b-4c86-982d-4e9efde21884/volumes" Dec 06 11:30:07 crc kubenswrapper[4678]: I1206 11:30:07.747688 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-dnkx5" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="registry-server" probeResult="failure" output=< Dec 06 11:30:07 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:30:07 crc kubenswrapper[4678]: > Dec 06 11:30:07 crc kubenswrapper[4678]: I1206 11:30:07.952315 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-68z4l" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="registry-server" probeResult="failure" output=< Dec 06 11:30:07 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:30:07 crc kubenswrapper[4678]: > Dec 06 11:30:14 crc kubenswrapper[4678]: I1206 11:30:14.356001 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:30:14 crc kubenswrapper[4678]: I1206 11:30:14.403089 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:30:14 crc kubenswrapper[4678]: I1206 11:30:14.599781 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sznb"] Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.019570 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7sznb" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="registry-server" containerID="cri-o://4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc" gracePeriod=2 Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.794657 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.824189 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.839382 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-utilities\") pod \"e407ea61-547c-43f5-842e-dad0c4597c07\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.839427 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc9nl\" (UniqueName: \"kubernetes.io/projected/e407ea61-547c-43f5-842e-dad0c4597c07-kube-api-access-wc9nl\") pod \"e407ea61-547c-43f5-842e-dad0c4597c07\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.839472 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-catalog-content\") pod \"e407ea61-547c-43f5-842e-dad0c4597c07\" (UID: \"e407ea61-547c-43f5-842e-dad0c4597c07\") " Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.846639 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-utilities" (OuterVolumeSpecName: "utilities") pod "e407ea61-547c-43f5-842e-dad0c4597c07" (UID: "e407ea61-547c-43f5-842e-dad0c4597c07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.871871 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e407ea61-547c-43f5-842e-dad0c4597c07" (UID: "e407ea61-547c-43f5-842e-dad0c4597c07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.885090 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.892653 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e407ea61-547c-43f5-842e-dad0c4597c07-kube-api-access-wc9nl" (OuterVolumeSpecName: "kube-api-access-wc9nl") pod "e407ea61-547c-43f5-842e-dad0c4597c07" (UID: "e407ea61-547c-43f5-842e-dad0c4597c07"). InnerVolumeSpecName "kube-api-access-wc9nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.941820 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.941858 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc9nl\" (UniqueName: \"kubernetes.io/projected/e407ea61-547c-43f5-842e-dad0c4597c07-kube-api-access-wc9nl\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:16 crc kubenswrapper[4678]: I1206 11:30:16.941870 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e407ea61-547c-43f5-842e-dad0c4597c07-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.032061 4678 generic.go:334] "Generic (PLEG): container finished" podID="e407ea61-547c-43f5-842e-dad0c4597c07" containerID="4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc" exitCode=0 Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.032162 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sznb" event={"ID":"e407ea61-547c-43f5-842e-dad0c4597c07","Type":"ContainerDied","Data":"4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc"} Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.032213 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7sznb" event={"ID":"e407ea61-547c-43f5-842e-dad0c4597c07","Type":"ContainerDied","Data":"09ff5aa3982dfe1f1082ec3daff093205f93699b3f4ea7121ba34a7f647ce869"} Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.032238 4678 scope.go:117] "RemoveContainer" containerID="4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.032431 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7sznb" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.060733 4678 scope.go:117] "RemoveContainer" containerID="c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.081154 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sznb"] Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.092417 4678 scope.go:117] "RemoveContainer" containerID="abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.094290 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7sznb"] Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.137617 4678 scope.go:117] "RemoveContainer" containerID="4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc" Dec 06 11:30:17 crc kubenswrapper[4678]: E1206 11:30:17.138140 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc\": container with ID starting with 4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc not found: ID does not exist" containerID="4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.138279 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc"} err="failed to get container status \"4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc\": rpc error: code = NotFound desc = could not find container \"4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc\": container with ID starting with 4c2c6f0506fcb0e0429f50945e623ccc488e740e9891b2b1abf99c68c7e2eebc not found: ID does not exist" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.138610 4678 scope.go:117] "RemoveContainer" containerID="c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03" Dec 06 11:30:17 crc kubenswrapper[4678]: E1206 11:30:17.139138 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03\": container with ID starting with c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03 not found: ID does not exist" containerID="c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.139191 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03"} err="failed to get container status \"c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03\": rpc error: code = NotFound desc = could not find container \"c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03\": container with ID starting with c12031d462ee4bdc5781b228ca1dca6d88ef8878af7d266ac9267f4d881dfe03 not found: ID does not exist" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.139239 4678 scope.go:117] "RemoveContainer" containerID="abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f" Dec 06 11:30:17 crc kubenswrapper[4678]: E1206 11:30:17.139662 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f\": container with ID starting with abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f not found: ID does not exist" containerID="abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.139691 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f"} err="failed to get container status \"abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f\": rpc error: code = NotFound desc = could not find container \"abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f\": container with ID starting with abad168a68bc8887d8435a8cdf87807e1d443f3c721547b2cae44f0f5632920f not found: ID does not exist" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.477580 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:30:17 crc kubenswrapper[4678]: E1206 11:30:17.478232 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.491370 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" path="/var/lib/kubelet/pods/e407ea61-547c-43f5-842e-dad0c4597c07/volumes" Dec 06 11:30:17 crc kubenswrapper[4678]: I1206 11:30:17.948644 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-68z4l" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="registry-server" probeResult="failure" output=< Dec 06 11:30:17 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:30:17 crc kubenswrapper[4678]: > Dec 06 11:30:18 crc kubenswrapper[4678]: I1206 11:30:18.802541 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dnkx5"] Dec 06 11:30:18 crc kubenswrapper[4678]: I1206 11:30:18.802974 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dnkx5" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="registry-server" containerID="cri-o://9e7a91291cf12f4aa21196898ca89b78353e117c5c842d660b0a692ccf55c085" gracePeriod=2 Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.052313 4678 generic.go:334] "Generic (PLEG): container finished" podID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerID="9e7a91291cf12f4aa21196898ca89b78353e117c5c842d660b0a692ccf55c085" exitCode=0 Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.052575 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnkx5" event={"ID":"d17a6604-cafe-4e6d-9df5-cb6ebc12720b","Type":"ContainerDied","Data":"9e7a91291cf12f4aa21196898ca89b78353e117c5c842d660b0a692ccf55c085"} Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.450542 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.497802 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rb7pz\" (UniqueName: \"kubernetes.io/projected/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-kube-api-access-rb7pz\") pod \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.500018 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-utilities\") pod \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.500080 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-catalog-content\") pod \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\" (UID: \"d17a6604-cafe-4e6d-9df5-cb6ebc12720b\") " Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.510669 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-utilities" (OuterVolumeSpecName: "utilities") pod "d17a6604-cafe-4e6d-9df5-cb6ebc12720b" (UID: "d17a6604-cafe-4e6d-9df5-cb6ebc12720b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.540777 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-kube-api-access-rb7pz" (OuterVolumeSpecName: "kube-api-access-rb7pz") pod "d17a6604-cafe-4e6d-9df5-cb6ebc12720b" (UID: "d17a6604-cafe-4e6d-9df5-cb6ebc12720b"). InnerVolumeSpecName "kube-api-access-rb7pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.602239 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d17a6604-cafe-4e6d-9df5-cb6ebc12720b" (UID: "d17a6604-cafe-4e6d-9df5-cb6ebc12720b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.609114 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rb7pz\" (UniqueName: \"kubernetes.io/projected/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-kube-api-access-rb7pz\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.609153 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:19 crc kubenswrapper[4678]: I1206 11:30:19.609166 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d17a6604-cafe-4e6d-9df5-cb6ebc12720b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:20 crc kubenswrapper[4678]: I1206 11:30:20.063441 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dnkx5" event={"ID":"d17a6604-cafe-4e6d-9df5-cb6ebc12720b","Type":"ContainerDied","Data":"71ed2fe114bf9a28e50078a503cbc9cef34fc0b54dbe7870a0a8a73750b74ae9"} Dec 06 11:30:20 crc kubenswrapper[4678]: I1206 11:30:20.063521 4678 scope.go:117] "RemoveContainer" containerID="9e7a91291cf12f4aa21196898ca89b78353e117c5c842d660b0a692ccf55c085" Dec 06 11:30:20 crc kubenswrapper[4678]: I1206 11:30:20.063578 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dnkx5" Dec 06 11:30:20 crc kubenswrapper[4678]: I1206 11:30:20.091713 4678 scope.go:117] "RemoveContainer" containerID="1d221d2c6f39c10763810959a570233ffb787caf1ea18646a79fcebdd590f954" Dec 06 11:30:20 crc kubenswrapper[4678]: I1206 11:30:20.114714 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dnkx5"] Dec 06 11:30:20 crc kubenswrapper[4678]: I1206 11:30:20.123701 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dnkx5"] Dec 06 11:30:20 crc kubenswrapper[4678]: I1206 11:30:20.141457 4678 scope.go:117] "RemoveContainer" containerID="52bb33290e0bb31d0522024a3bcc2e3a1df93f070df94df0133bbcaa7221c3bd" Dec 06 11:30:21 crc kubenswrapper[4678]: I1206 11:30:21.487315 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" path="/var/lib/kubelet/pods/d17a6604-cafe-4e6d-9df5-cb6ebc12720b/volumes" Dec 06 11:30:26 crc kubenswrapper[4678]: I1206 11:30:26.951454 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:30:27 crc kubenswrapper[4678]: I1206 11:30:27.010375 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:30:27 crc kubenswrapper[4678]: I1206 11:30:27.576122 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68z4l"] Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.136923 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-68z4l" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="registry-server" containerID="cri-o://c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb" gracePeriod=2 Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.776057 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.877304 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-catalog-content\") pod \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.877368 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfn8h\" (UniqueName: \"kubernetes.io/projected/4aa99324-976f-4fbe-9c05-edf1f0ff783a-kube-api-access-kfn8h\") pod \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.877391 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-utilities\") pod \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\" (UID: \"4aa99324-976f-4fbe-9c05-edf1f0ff783a\") " Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.878355 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-utilities" (OuterVolumeSpecName: "utilities") pod "4aa99324-976f-4fbe-9c05-edf1f0ff783a" (UID: "4aa99324-976f-4fbe-9c05-edf1f0ff783a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.882887 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4aa99324-976f-4fbe-9c05-edf1f0ff783a-kube-api-access-kfn8h" (OuterVolumeSpecName: "kube-api-access-kfn8h") pod "4aa99324-976f-4fbe-9c05-edf1f0ff783a" (UID: "4aa99324-976f-4fbe-9c05-edf1f0ff783a"). InnerVolumeSpecName "kube-api-access-kfn8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.979443 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfn8h\" (UniqueName: \"kubernetes.io/projected/4aa99324-976f-4fbe-9c05-edf1f0ff783a-kube-api-access-kfn8h\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.979778 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:28 crc kubenswrapper[4678]: I1206 11:30:28.990765 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4aa99324-976f-4fbe-9c05-edf1f0ff783a" (UID: "4aa99324-976f-4fbe-9c05-edf1f0ff783a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.081053 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4aa99324-976f-4fbe-9c05-edf1f0ff783a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.148154 4678 generic.go:334] "Generic (PLEG): container finished" podID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerID="c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb" exitCode=0 Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.148363 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68z4l" event={"ID":"4aa99324-976f-4fbe-9c05-edf1f0ff783a","Type":"ContainerDied","Data":"c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb"} Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.149248 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-68z4l" event={"ID":"4aa99324-976f-4fbe-9c05-edf1f0ff783a","Type":"ContainerDied","Data":"f819730d2fa67f6a6aa1832c0ef0faa7e2643ea02d9bf23bfbf206b7d180bad0"} Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.149357 4678 scope.go:117] "RemoveContainer" containerID="c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.148449 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-68z4l" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.168808 4678 scope.go:117] "RemoveContainer" containerID="777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.196758 4678 scope.go:117] "RemoveContainer" containerID="238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.196761 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-68z4l"] Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.208824 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-68z4l"] Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.240916 4678 scope.go:117] "RemoveContainer" containerID="c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb" Dec 06 11:30:29 crc kubenswrapper[4678]: E1206 11:30:29.241374 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb\": container with ID starting with c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb not found: ID does not exist" containerID="c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.241422 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb"} err="failed to get container status \"c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb\": rpc error: code = NotFound desc = could not find container \"c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb\": container with ID starting with c8140b1dd6a1753265e8f9a3c1acb11b51dfe9457ce738c7e05e9c6e7eb1b4fb not found: ID does not exist" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.241449 4678 scope.go:117] "RemoveContainer" containerID="777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd" Dec 06 11:30:29 crc kubenswrapper[4678]: E1206 11:30:29.241896 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd\": container with ID starting with 777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd not found: ID does not exist" containerID="777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.241942 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd"} err="failed to get container status \"777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd\": rpc error: code = NotFound desc = could not find container \"777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd\": container with ID starting with 777ad4f9af70ebf4a81b12271129f4d7baf98480052ddf06f926320428ed62cd not found: ID does not exist" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.241967 4678 scope.go:117] "RemoveContainer" containerID="238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd" Dec 06 11:30:29 crc kubenswrapper[4678]: E1206 11:30:29.242247 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd\": container with ID starting with 238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd not found: ID does not exist" containerID="238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.242272 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd"} err="failed to get container status \"238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd\": rpc error: code = NotFound desc = could not find container \"238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd\": container with ID starting with 238149a417e8bb7a5d0fd7587ef26a49db91ac93b537c24f6e34201db5f38afd not found: ID does not exist" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.476708 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:30:29 crc kubenswrapper[4678]: E1206 11:30:29.477008 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:30:29 crc kubenswrapper[4678]: I1206 11:30:29.490731 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" path="/var/lib/kubelet/pods/4aa99324-976f-4fbe-9c05-edf1f0ff783a/volumes" Dec 06 11:30:43 crc kubenswrapper[4678]: I1206 11:30:43.477271 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:30:43 crc kubenswrapper[4678]: E1206 11:30:43.478226 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:30:48 crc kubenswrapper[4678]: I1206 11:30:48.118472 4678 scope.go:117] "RemoveContainer" containerID="02fcb2fdee791cd878276c1779be8ef14f72286bfb3c0db5717ea1f2e12ba311" Dec 06 11:30:56 crc kubenswrapper[4678]: I1206 11:30:56.476619 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:30:56 crc kubenswrapper[4678]: E1206 11:30:56.477175 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:31:07 crc kubenswrapper[4678]: I1206 11:31:07.625006 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:31:08 crc kubenswrapper[4678]: I1206 11:31:08.682883 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"6ccbaed23006ca0948e15c55fb29f77ca16e887730b6731fd47f6d3072b6dd41"} Dec 06 11:33:29 crc kubenswrapper[4678]: I1206 11:33:29.505312 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:33:29 crc kubenswrapper[4678]: I1206 11:33:29.505959 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:33:59 crc kubenswrapper[4678]: I1206 11:33:59.505307 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:33:59 crc kubenswrapper[4678]: I1206 11:33:59.506614 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:34:29 crc kubenswrapper[4678]: I1206 11:34:29.506667 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:34:29 crc kubenswrapper[4678]: I1206 11:34:29.508263 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:34:29 crc kubenswrapper[4678]: I1206 11:34:29.508384 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:34:29 crc kubenswrapper[4678]: I1206 11:34:29.509198 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ccbaed23006ca0948e15c55fb29f77ca16e887730b6731fd47f6d3072b6dd41"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:34:29 crc kubenswrapper[4678]: I1206 11:34:29.509339 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://6ccbaed23006ca0948e15c55fb29f77ca16e887730b6731fd47f6d3072b6dd41" gracePeriod=600 Dec 06 11:34:30 crc kubenswrapper[4678]: I1206 11:34:30.492345 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="6ccbaed23006ca0948e15c55fb29f77ca16e887730b6731fd47f6d3072b6dd41" exitCode=0 Dec 06 11:34:30 crc kubenswrapper[4678]: I1206 11:34:30.492384 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"6ccbaed23006ca0948e15c55fb29f77ca16e887730b6731fd47f6d3072b6dd41"} Dec 06 11:34:30 crc kubenswrapper[4678]: I1206 11:34:30.492986 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1"} Dec 06 11:34:30 crc kubenswrapper[4678]: I1206 11:34:30.493009 4678 scope.go:117] "RemoveContainer" containerID="ead748ef15bf1953311ec036d087d289a9d69e0d275f57a0ae8694e4ed9d4a73" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.843638 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ws9zt"] Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844433 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="extract-content" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844449 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="extract-content" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844467 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="extract-content" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844475 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="extract-content" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844501 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="extract-utilities" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844511 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="extract-utilities" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844531 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844538 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844549 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b175a69-f992-4f15-8257-a9a91b1b364f" containerName="collect-profiles" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844556 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b175a69-f992-4f15-8257-a9a91b1b364f" containerName="collect-profiles" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844570 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844576 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844590 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="extract-utilities" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844597 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="extract-utilities" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844616 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="extract-utilities" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844623 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="extract-utilities" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844632 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="extract-content" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844639 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="extract-content" Dec 06 11:34:33 crc kubenswrapper[4678]: E1206 11:34:33.844659 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844665 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844870 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="4aa99324-976f-4fbe-9c05-edf1f0ff783a" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.844980 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b175a69-f992-4f15-8257-a9a91b1b364f" containerName="collect-profiles" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.845001 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17a6604-cafe-4e6d-9df5-cb6ebc12720b" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.845014 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e407ea61-547c-43f5-842e-dad0c4597c07" containerName="registry-server" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.846660 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.865328 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ws9zt"] Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.992449 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5ttk\" (UniqueName: \"kubernetes.io/projected/431f1571-d5ac-477e-870d-0e2c28241fda-kube-api-access-l5ttk\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.992714 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-catalog-content\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:33 crc kubenswrapper[4678]: I1206 11:34:33.992766 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-utilities\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.094679 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5ttk\" (UniqueName: \"kubernetes.io/projected/431f1571-d5ac-477e-870d-0e2c28241fda-kube-api-access-l5ttk\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.094876 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-catalog-content\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.094945 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-utilities\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.095358 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-catalog-content\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.095451 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-utilities\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.132639 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5ttk\" (UniqueName: \"kubernetes.io/projected/431f1571-d5ac-477e-870d-0e2c28241fda-kube-api-access-l5ttk\") pod \"certified-operators-ws9zt\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.170660 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:34 crc kubenswrapper[4678]: I1206 11:34:34.735321 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ws9zt"] Dec 06 11:34:35 crc kubenswrapper[4678]: I1206 11:34:35.556965 4678 generic.go:334] "Generic (PLEG): container finished" podID="431f1571-d5ac-477e-870d-0e2c28241fda" containerID="bed63672978a4012e912f0abc9615c79406e8da29c0315a1b68cdbfa96b1e63f" exitCode=0 Dec 06 11:34:35 crc kubenswrapper[4678]: I1206 11:34:35.557047 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws9zt" event={"ID":"431f1571-d5ac-477e-870d-0e2c28241fda","Type":"ContainerDied","Data":"bed63672978a4012e912f0abc9615c79406e8da29c0315a1b68cdbfa96b1e63f"} Dec 06 11:34:35 crc kubenswrapper[4678]: I1206 11:34:35.557613 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws9zt" event={"ID":"431f1571-d5ac-477e-870d-0e2c28241fda","Type":"ContainerStarted","Data":"ab66e0afee50eb0c546bd5479976b5cc9524fa2d071e7edc90acddff873523d6"} Dec 06 11:34:36 crc kubenswrapper[4678]: I1206 11:34:36.577759 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws9zt" event={"ID":"431f1571-d5ac-477e-870d-0e2c28241fda","Type":"ContainerStarted","Data":"e241b9824268f726ecc36cc1623e1501883e5cb3fb43eafb635112e9de8ac0a9"} Dec 06 11:34:37 crc kubenswrapper[4678]: I1206 11:34:37.589346 4678 generic.go:334] "Generic (PLEG): container finished" podID="431f1571-d5ac-477e-870d-0e2c28241fda" containerID="e241b9824268f726ecc36cc1623e1501883e5cb3fb43eafb635112e9de8ac0a9" exitCode=0 Dec 06 11:34:37 crc kubenswrapper[4678]: I1206 11:34:37.589780 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws9zt" event={"ID":"431f1571-d5ac-477e-870d-0e2c28241fda","Type":"ContainerDied","Data":"e241b9824268f726ecc36cc1623e1501883e5cb3fb43eafb635112e9de8ac0a9"} Dec 06 11:34:38 crc kubenswrapper[4678]: I1206 11:34:38.602105 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws9zt" event={"ID":"431f1571-d5ac-477e-870d-0e2c28241fda","Type":"ContainerStarted","Data":"892569cf3810e50efc0465b059e5c4d754293c301fda567356dd6c99d658a45b"} Dec 06 11:34:44 crc kubenswrapper[4678]: I1206 11:34:44.171941 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:44 crc kubenswrapper[4678]: I1206 11:34:44.173269 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:44 crc kubenswrapper[4678]: I1206 11:34:44.237645 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:44 crc kubenswrapper[4678]: I1206 11:34:44.256662 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ws9zt" podStartSLOduration=8.772035054 podStartE2EDuration="11.256640094s" podCreationTimestamp="2025-12-06 11:34:33 +0000 UTC" firstStartedPulling="2025-12-06 11:34:35.560189665 +0000 UTC m=+3480.403621104" lastFinishedPulling="2025-12-06 11:34:38.044794705 +0000 UTC m=+3482.888226144" observedRunningTime="2025-12-06 11:34:38.62608811 +0000 UTC m=+3483.469519559" watchObservedRunningTime="2025-12-06 11:34:44.256640094 +0000 UTC m=+3489.100071533" Dec 06 11:34:44 crc kubenswrapper[4678]: I1206 11:34:44.723785 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:44 crc kubenswrapper[4678]: I1206 11:34:44.772512 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ws9zt"] Dec 06 11:34:46 crc kubenswrapper[4678]: I1206 11:34:46.691851 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ws9zt" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="registry-server" containerID="cri-o://892569cf3810e50efc0465b059e5c4d754293c301fda567356dd6c99d658a45b" gracePeriod=2 Dec 06 11:34:46 crc kubenswrapper[4678]: E1206 11:34:46.967624 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod431f1571_d5ac_477e_870d_0e2c28241fda.slice/crio-conmon-892569cf3810e50efc0465b059e5c4d754293c301fda567356dd6c99d658a45b.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.704514 4678 generic.go:334] "Generic (PLEG): container finished" podID="431f1571-d5ac-477e-870d-0e2c28241fda" containerID="892569cf3810e50efc0465b059e5c4d754293c301fda567356dd6c99d658a45b" exitCode=0 Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.704661 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws9zt" event={"ID":"431f1571-d5ac-477e-870d-0e2c28241fda","Type":"ContainerDied","Data":"892569cf3810e50efc0465b059e5c4d754293c301fda567356dd6c99d658a45b"} Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.896958 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.947716 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-catalog-content\") pod \"431f1571-d5ac-477e-870d-0e2c28241fda\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.947792 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5ttk\" (UniqueName: \"kubernetes.io/projected/431f1571-d5ac-477e-870d-0e2c28241fda-kube-api-access-l5ttk\") pod \"431f1571-d5ac-477e-870d-0e2c28241fda\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.947889 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-utilities\") pod \"431f1571-d5ac-477e-870d-0e2c28241fda\" (UID: \"431f1571-d5ac-477e-870d-0e2c28241fda\") " Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.948790 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-utilities" (OuterVolumeSpecName: "utilities") pod "431f1571-d5ac-477e-870d-0e2c28241fda" (UID: "431f1571-d5ac-477e-870d-0e2c28241fda"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:34:47 crc kubenswrapper[4678]: I1206 11:34:47.953532 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/431f1571-d5ac-477e-870d-0e2c28241fda-kube-api-access-l5ttk" (OuterVolumeSpecName: "kube-api-access-l5ttk") pod "431f1571-d5ac-477e-870d-0e2c28241fda" (UID: "431f1571-d5ac-477e-870d-0e2c28241fda"). InnerVolumeSpecName "kube-api-access-l5ttk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.000525 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "431f1571-d5ac-477e-870d-0e2c28241fda" (UID: "431f1571-d5ac-477e-870d-0e2c28241fda"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.049870 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.049908 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5ttk\" (UniqueName: \"kubernetes.io/projected/431f1571-d5ac-477e-870d-0e2c28241fda-kube-api-access-l5ttk\") on node \"crc\" DevicePath \"\"" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.049922 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431f1571-d5ac-477e-870d-0e2c28241fda-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.717304 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ws9zt" event={"ID":"431f1571-d5ac-477e-870d-0e2c28241fda","Type":"ContainerDied","Data":"ab66e0afee50eb0c546bd5479976b5cc9524fa2d071e7edc90acddff873523d6"} Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.717376 4678 scope.go:117] "RemoveContainer" containerID="892569cf3810e50efc0465b059e5c4d754293c301fda567356dd6c99d658a45b" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.717394 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ws9zt" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.745968 4678 scope.go:117] "RemoveContainer" containerID="e241b9824268f726ecc36cc1623e1501883e5cb3fb43eafb635112e9de8ac0a9" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.776193 4678 scope.go:117] "RemoveContainer" containerID="bed63672978a4012e912f0abc9615c79406e8da29c0315a1b68cdbfa96b1e63f" Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.818605 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ws9zt"] Dec 06 11:34:48 crc kubenswrapper[4678]: I1206 11:34:48.834088 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ws9zt"] Dec 06 11:34:49 crc kubenswrapper[4678]: I1206 11:34:49.486392 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" path="/var/lib/kubelet/pods/431f1571-d5ac-477e-870d-0e2c28241fda/volumes" Dec 06 11:36:29 crc kubenswrapper[4678]: I1206 11:36:29.505144 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:36:29 crc kubenswrapper[4678]: I1206 11:36:29.506779 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:36:59 crc kubenswrapper[4678]: I1206 11:36:59.505634 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:36:59 crc kubenswrapper[4678]: I1206 11:36:59.506211 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:37:29 crc kubenswrapper[4678]: I1206 11:37:29.505463 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:37:29 crc kubenswrapper[4678]: I1206 11:37:29.505978 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:37:29 crc kubenswrapper[4678]: I1206 11:37:29.506023 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:37:29 crc kubenswrapper[4678]: I1206 11:37:29.506754 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:37:29 crc kubenswrapper[4678]: I1206 11:37:29.506803 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" gracePeriod=600 Dec 06 11:37:29 crc kubenswrapper[4678]: E1206 11:37:29.632658 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:37:30 crc kubenswrapper[4678]: I1206 11:37:30.628651 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" exitCode=0 Dec 06 11:37:30 crc kubenswrapper[4678]: I1206 11:37:30.628683 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1"} Dec 06 11:37:30 crc kubenswrapper[4678]: I1206 11:37:30.628941 4678 scope.go:117] "RemoveContainer" containerID="6ccbaed23006ca0948e15c55fb29f77ca16e887730b6731fd47f6d3072b6dd41" Dec 06 11:37:30 crc kubenswrapper[4678]: I1206 11:37:30.629535 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:37:30 crc kubenswrapper[4678]: E1206 11:37:30.629791 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:37:45 crc kubenswrapper[4678]: I1206 11:37:45.484019 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:37:45 crc kubenswrapper[4678]: E1206 11:37:45.484834 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:37:57 crc kubenswrapper[4678]: I1206 11:37:57.475919 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:37:57 crc kubenswrapper[4678]: E1206 11:37:57.476776 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:38:12 crc kubenswrapper[4678]: I1206 11:38:12.476408 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:38:12 crc kubenswrapper[4678]: E1206 11:38:12.477290 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:38:23 crc kubenswrapper[4678]: I1206 11:38:23.476582 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:38:23 crc kubenswrapper[4678]: E1206 11:38:23.477275 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:38:37 crc kubenswrapper[4678]: I1206 11:38:37.475500 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:38:37 crc kubenswrapper[4678]: E1206 11:38:37.476327 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:38:52 crc kubenswrapper[4678]: I1206 11:38:52.478352 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:38:52 crc kubenswrapper[4678]: E1206 11:38:52.479974 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:39:04 crc kubenswrapper[4678]: I1206 11:39:04.475926 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:39:04 crc kubenswrapper[4678]: E1206 11:39:04.476724 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:39:18 crc kubenswrapper[4678]: I1206 11:39:18.476135 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:39:18 crc kubenswrapper[4678]: E1206 11:39:18.477064 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:39:29 crc kubenswrapper[4678]: I1206 11:39:29.476475 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:39:29 crc kubenswrapper[4678]: E1206 11:39:29.477213 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:39:42 crc kubenswrapper[4678]: I1206 11:39:42.476293 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:39:42 crc kubenswrapper[4678]: E1206 11:39:42.477291 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:39:54 crc kubenswrapper[4678]: I1206 11:39:54.476346 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:39:54 crc kubenswrapper[4678]: E1206 11:39:54.477137 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:40:05 crc kubenswrapper[4678]: I1206 11:40:05.489617 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:40:05 crc kubenswrapper[4678]: E1206 11:40:05.492087 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:40:18 crc kubenswrapper[4678]: I1206 11:40:18.476080 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:40:18 crc kubenswrapper[4678]: E1206 11:40:18.476738 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.240610 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rsrhk"] Dec 06 11:40:22 crc kubenswrapper[4678]: E1206 11:40:22.241473 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="registry-server" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.241500 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="registry-server" Dec 06 11:40:22 crc kubenswrapper[4678]: E1206 11:40:22.241524 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="extract-content" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.241531 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="extract-content" Dec 06 11:40:22 crc kubenswrapper[4678]: E1206 11:40:22.241549 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="extract-utilities" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.241555 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="extract-utilities" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.241742 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="431f1571-d5ac-477e-870d-0e2c28241fda" containerName="registry-server" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.243018 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.258235 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsrhk"] Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.397261 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm6k6\" (UniqueName: \"kubernetes.io/projected/9a70063b-28fa-4ffc-88c2-29d790e853e6-kube-api-access-nm6k6\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.397330 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-catalog-content\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.397406 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-utilities\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.499469 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm6k6\" (UniqueName: \"kubernetes.io/projected/9a70063b-28fa-4ffc-88c2-29d790e853e6-kube-api-access-nm6k6\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.499571 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-catalog-content\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.499649 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-utilities\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.500057 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-catalog-content\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.500089 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-utilities\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.531810 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm6k6\" (UniqueName: \"kubernetes.io/projected/9a70063b-28fa-4ffc-88c2-29d790e853e6-kube-api-access-nm6k6\") pod \"community-operators-rsrhk\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:22 crc kubenswrapper[4678]: I1206 11:40:22.577164 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:23 crc kubenswrapper[4678]: I1206 11:40:23.190414 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rsrhk"] Dec 06 11:40:23 crc kubenswrapper[4678]: I1206 11:40:23.205311 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsrhk" event={"ID":"9a70063b-28fa-4ffc-88c2-29d790e853e6","Type":"ContainerStarted","Data":"c3bffd055edf3b5a6c359fadd3c08b97169e6eeb80251c2c52cec861e84fcc1d"} Dec 06 11:40:24 crc kubenswrapper[4678]: I1206 11:40:24.217067 4678 generic.go:334] "Generic (PLEG): container finished" podID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerID="eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2" exitCode=0 Dec 06 11:40:24 crc kubenswrapper[4678]: I1206 11:40:24.217173 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsrhk" event={"ID":"9a70063b-28fa-4ffc-88c2-29d790e853e6","Type":"ContainerDied","Data":"eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2"} Dec 06 11:40:24 crc kubenswrapper[4678]: I1206 11:40:24.219548 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:40:25 crc kubenswrapper[4678]: I1206 11:40:25.239342 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsrhk" event={"ID":"9a70063b-28fa-4ffc-88c2-29d790e853e6","Type":"ContainerStarted","Data":"91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931"} Dec 06 11:40:26 crc kubenswrapper[4678]: I1206 11:40:26.251289 4678 generic.go:334] "Generic (PLEG): container finished" podID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerID="91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931" exitCode=0 Dec 06 11:40:26 crc kubenswrapper[4678]: I1206 11:40:26.251357 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsrhk" event={"ID":"9a70063b-28fa-4ffc-88c2-29d790e853e6","Type":"ContainerDied","Data":"91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931"} Dec 06 11:40:27 crc kubenswrapper[4678]: I1206 11:40:27.269321 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsrhk" event={"ID":"9a70063b-28fa-4ffc-88c2-29d790e853e6","Type":"ContainerStarted","Data":"5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79"} Dec 06 11:40:31 crc kubenswrapper[4678]: I1206 11:40:31.482797 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:40:31 crc kubenswrapper[4678]: E1206 11:40:31.483678 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:40:32 crc kubenswrapper[4678]: I1206 11:40:32.578129 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:32 crc kubenswrapper[4678]: I1206 11:40:32.578349 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:32 crc kubenswrapper[4678]: I1206 11:40:32.625074 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:32 crc kubenswrapper[4678]: I1206 11:40:32.657615 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rsrhk" podStartSLOduration=8.172226508 podStartE2EDuration="10.65758005s" podCreationTimestamp="2025-12-06 11:40:22 +0000 UTC" firstStartedPulling="2025-12-06 11:40:24.21923815 +0000 UTC m=+3829.062669599" lastFinishedPulling="2025-12-06 11:40:26.704591702 +0000 UTC m=+3831.548023141" observedRunningTime="2025-12-06 11:40:27.303950084 +0000 UTC m=+3832.147381523" watchObservedRunningTime="2025-12-06 11:40:32.65758005 +0000 UTC m=+3837.501011499" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.171018 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wpmf9"] Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.173731 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.230566 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wpmf9"] Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.321329 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcmcg\" (UniqueName: \"kubernetes.io/projected/af22b429-f8e3-45cf-af1f-b37660da92c4-kube-api-access-vcmcg\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.321479 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-utilities\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.321728 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-catalog-content\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.376129 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.423916 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcmcg\" (UniqueName: \"kubernetes.io/projected/af22b429-f8e3-45cf-af1f-b37660da92c4-kube-api-access-vcmcg\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.424057 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-utilities\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.424113 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-catalog-content\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.424778 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-catalog-content\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.424846 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-utilities\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.442455 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcmcg\" (UniqueName: \"kubernetes.io/projected/af22b429-f8e3-45cf-af1f-b37660da92c4-kube-api-access-vcmcg\") pod \"redhat-marketplace-wpmf9\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:33 crc kubenswrapper[4678]: I1206 11:40:33.497185 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:34 crc kubenswrapper[4678]: I1206 11:40:34.021411 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wpmf9"] Dec 06 11:40:34 crc kubenswrapper[4678]: I1206 11:40:34.331835 4678 generic.go:334] "Generic (PLEG): container finished" podID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerID="0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033" exitCode=0 Dec 06 11:40:34 crc kubenswrapper[4678]: I1206 11:40:34.332647 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wpmf9" event={"ID":"af22b429-f8e3-45cf-af1f-b37660da92c4","Type":"ContainerDied","Data":"0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033"} Dec 06 11:40:34 crc kubenswrapper[4678]: I1206 11:40:34.332684 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wpmf9" event={"ID":"af22b429-f8e3-45cf-af1f-b37660da92c4","Type":"ContainerStarted","Data":"9874f774c6c03feeb0a0267ccf5f40bebf85acdf040d11da6bd8abfbbfff96b9"} Dec 06 11:40:35 crc kubenswrapper[4678]: I1206 11:40:35.345936 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wpmf9" event={"ID":"af22b429-f8e3-45cf-af1f-b37660da92c4","Type":"ContainerStarted","Data":"5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e"} Dec 06 11:40:35 crc kubenswrapper[4678]: I1206 11:40:35.673468 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsrhk"] Dec 06 11:40:35 crc kubenswrapper[4678]: I1206 11:40:35.674032 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rsrhk" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="registry-server" containerID="cri-o://5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79" gracePeriod=2 Dec 06 11:40:35 crc kubenswrapper[4678]: E1206 11:40:35.904717 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf22b429_f8e3_45cf_af1f_b37660da92c4.slice/crio-5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a70063b_28fa_4ffc_88c2_29d790e853e6.slice/crio-5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a70063b_28fa_4ffc_88c2_29d790e853e6.slice/crio-conmon-5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.221318 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.362766 4678 generic.go:334] "Generic (PLEG): container finished" podID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerID="5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e" exitCode=0 Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.362868 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wpmf9" event={"ID":"af22b429-f8e3-45cf-af1f-b37660da92c4","Type":"ContainerDied","Data":"5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e"} Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.372718 4678 generic.go:334] "Generic (PLEG): container finished" podID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerID="5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79" exitCode=0 Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.372757 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsrhk" event={"ID":"9a70063b-28fa-4ffc-88c2-29d790e853e6","Type":"ContainerDied","Data":"5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79"} Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.372783 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rsrhk" event={"ID":"9a70063b-28fa-4ffc-88c2-29d790e853e6","Type":"ContainerDied","Data":"c3bffd055edf3b5a6c359fadd3c08b97169e6eeb80251c2c52cec861e84fcc1d"} Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.372800 4678 scope.go:117] "RemoveContainer" containerID="5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.373084 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rsrhk" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.404364 4678 scope.go:117] "RemoveContainer" containerID="91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.421629 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-utilities\") pod \"9a70063b-28fa-4ffc-88c2-29d790e853e6\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.421732 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm6k6\" (UniqueName: \"kubernetes.io/projected/9a70063b-28fa-4ffc-88c2-29d790e853e6-kube-api-access-nm6k6\") pod \"9a70063b-28fa-4ffc-88c2-29d790e853e6\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.421871 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-catalog-content\") pod \"9a70063b-28fa-4ffc-88c2-29d790e853e6\" (UID: \"9a70063b-28fa-4ffc-88c2-29d790e853e6\") " Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.422393 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-utilities" (OuterVolumeSpecName: "utilities") pod "9a70063b-28fa-4ffc-88c2-29d790e853e6" (UID: "9a70063b-28fa-4ffc-88c2-29d790e853e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.427051 4678 scope.go:117] "RemoveContainer" containerID="eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.431827 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a70063b-28fa-4ffc-88c2-29d790e853e6-kube-api-access-nm6k6" (OuterVolumeSpecName: "kube-api-access-nm6k6") pod "9a70063b-28fa-4ffc-88c2-29d790e853e6" (UID: "9a70063b-28fa-4ffc-88c2-29d790e853e6"). InnerVolumeSpecName "kube-api-access-nm6k6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.485553 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a70063b-28fa-4ffc-88c2-29d790e853e6" (UID: "9a70063b-28fa-4ffc-88c2-29d790e853e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.508453 4678 scope.go:117] "RemoveContainer" containerID="5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79" Dec 06 11:40:36 crc kubenswrapper[4678]: E1206 11:40:36.509078 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79\": container with ID starting with 5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79 not found: ID does not exist" containerID="5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.509128 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79"} err="failed to get container status \"5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79\": rpc error: code = NotFound desc = could not find container \"5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79\": container with ID starting with 5158e229891b371c838c4fc457b2b8158d6c89ac860ac026c4307690cf38ff79 not found: ID does not exist" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.509179 4678 scope.go:117] "RemoveContainer" containerID="91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931" Dec 06 11:40:36 crc kubenswrapper[4678]: E1206 11:40:36.509505 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931\": container with ID starting with 91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931 not found: ID does not exist" containerID="91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.509534 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931"} err="failed to get container status \"91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931\": rpc error: code = NotFound desc = could not find container \"91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931\": container with ID starting with 91a527e1a1069e1bf683225ebf01f02bbee97b3e1239da0b4d5c8e319bb36931 not found: ID does not exist" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.509551 4678 scope.go:117] "RemoveContainer" containerID="eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2" Dec 06 11:40:36 crc kubenswrapper[4678]: E1206 11:40:36.509831 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2\": container with ID starting with eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2 not found: ID does not exist" containerID="eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.509885 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2"} err="failed to get container status \"eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2\": rpc error: code = NotFound desc = could not find container \"eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2\": container with ID starting with eea1ad75c07f4753222ed437688b0d815a7e75f7c2d2981ce0a71bf05ace49b2 not found: ID does not exist" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.524858 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.524997 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm6k6\" (UniqueName: \"kubernetes.io/projected/9a70063b-28fa-4ffc-88c2-29d790e853e6-kube-api-access-nm6k6\") on node \"crc\" DevicePath \"\"" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.525024 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a70063b-28fa-4ffc-88c2-29d790e853e6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.709334 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rsrhk"] Dec 06 11:40:36 crc kubenswrapper[4678]: I1206 11:40:36.719160 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rsrhk"] Dec 06 11:40:37 crc kubenswrapper[4678]: I1206 11:40:37.382969 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wpmf9" event={"ID":"af22b429-f8e3-45cf-af1f-b37660da92c4","Type":"ContainerStarted","Data":"e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6"} Dec 06 11:40:37 crc kubenswrapper[4678]: I1206 11:40:37.413649 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wpmf9" podStartSLOduration=2.00818375 podStartE2EDuration="4.413623469s" podCreationTimestamp="2025-12-06 11:40:33 +0000 UTC" firstStartedPulling="2025-12-06 11:40:34.334863431 +0000 UTC m=+3839.178294870" lastFinishedPulling="2025-12-06 11:40:36.74030315 +0000 UTC m=+3841.583734589" observedRunningTime="2025-12-06 11:40:37.408786628 +0000 UTC m=+3842.252218077" watchObservedRunningTime="2025-12-06 11:40:37.413623469 +0000 UTC m=+3842.257054908" Dec 06 11:40:37 crc kubenswrapper[4678]: I1206 11:40:37.487504 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" path="/var/lib/kubelet/pods/9a70063b-28fa-4ffc-88c2-29d790e853e6/volumes" Dec 06 11:40:43 crc kubenswrapper[4678]: I1206 11:40:43.477480 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:40:43 crc kubenswrapper[4678]: E1206 11:40:43.478263 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:40:43 crc kubenswrapper[4678]: I1206 11:40:43.497741 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:43 crc kubenswrapper[4678]: I1206 11:40:43.498157 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:43 crc kubenswrapper[4678]: I1206 11:40:43.548143 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:44 crc kubenswrapper[4678]: I1206 11:40:44.493155 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:44 crc kubenswrapper[4678]: I1206 11:40:44.544724 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wpmf9"] Dec 06 11:40:46 crc kubenswrapper[4678]: E1206 11:40:46.165761 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf22b429_f8e3_45cf_af1f_b37660da92c4.slice/crio-5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:40:46 crc kubenswrapper[4678]: I1206 11:40:46.463198 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wpmf9" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="registry-server" containerID="cri-o://e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6" gracePeriod=2 Dec 06 11:40:46 crc kubenswrapper[4678]: I1206 11:40:46.981160 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.104949 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcmcg\" (UniqueName: \"kubernetes.io/projected/af22b429-f8e3-45cf-af1f-b37660da92c4-kube-api-access-vcmcg\") pod \"af22b429-f8e3-45cf-af1f-b37660da92c4\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.105100 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-utilities\") pod \"af22b429-f8e3-45cf-af1f-b37660da92c4\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.105162 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-catalog-content\") pod \"af22b429-f8e3-45cf-af1f-b37660da92c4\" (UID: \"af22b429-f8e3-45cf-af1f-b37660da92c4\") " Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.106078 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-utilities" (OuterVolumeSpecName: "utilities") pod "af22b429-f8e3-45cf-af1f-b37660da92c4" (UID: "af22b429-f8e3-45cf-af1f-b37660da92c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.112723 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af22b429-f8e3-45cf-af1f-b37660da92c4-kube-api-access-vcmcg" (OuterVolumeSpecName: "kube-api-access-vcmcg") pod "af22b429-f8e3-45cf-af1f-b37660da92c4" (UID: "af22b429-f8e3-45cf-af1f-b37660da92c4"). InnerVolumeSpecName "kube-api-access-vcmcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.127151 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af22b429-f8e3-45cf-af1f-b37660da92c4" (UID: "af22b429-f8e3-45cf-af1f-b37660da92c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.207726 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcmcg\" (UniqueName: \"kubernetes.io/projected/af22b429-f8e3-45cf-af1f-b37660da92c4-kube-api-access-vcmcg\") on node \"crc\" DevicePath \"\"" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.207759 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.207770 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af22b429-f8e3-45cf-af1f-b37660da92c4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.481991 4678 generic.go:334] "Generic (PLEG): container finished" podID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerID="e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6" exitCode=0 Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.482082 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wpmf9" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.486982 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wpmf9" event={"ID":"af22b429-f8e3-45cf-af1f-b37660da92c4","Type":"ContainerDied","Data":"e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6"} Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.487188 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wpmf9" event={"ID":"af22b429-f8e3-45cf-af1f-b37660da92c4","Type":"ContainerDied","Data":"9874f774c6c03feeb0a0267ccf5f40bebf85acdf040d11da6bd8abfbbfff96b9"} Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.487277 4678 scope.go:117] "RemoveContainer" containerID="e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.520274 4678 scope.go:117] "RemoveContainer" containerID="5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.524355 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wpmf9"] Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.539753 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wpmf9"] Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.540285 4678 scope.go:117] "RemoveContainer" containerID="0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.601081 4678 scope.go:117] "RemoveContainer" containerID="e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6" Dec 06 11:40:47 crc kubenswrapper[4678]: E1206 11:40:47.601673 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6\": container with ID starting with e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6 not found: ID does not exist" containerID="e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.601702 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6"} err="failed to get container status \"e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6\": rpc error: code = NotFound desc = could not find container \"e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6\": container with ID starting with e63a62cba79efd9576dae4a474c5dc56778c2b61c3b4e6fe0ad9a27f8e6d87e6 not found: ID does not exist" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.601720 4678 scope.go:117] "RemoveContainer" containerID="5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e" Dec 06 11:40:47 crc kubenswrapper[4678]: E1206 11:40:47.602071 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e\": container with ID starting with 5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e not found: ID does not exist" containerID="5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.602088 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e"} err="failed to get container status \"5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e\": rpc error: code = NotFound desc = could not find container \"5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e\": container with ID starting with 5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e not found: ID does not exist" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.602100 4678 scope.go:117] "RemoveContainer" containerID="0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033" Dec 06 11:40:47 crc kubenswrapper[4678]: E1206 11:40:47.614262 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033\": container with ID starting with 0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033 not found: ID does not exist" containerID="0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033" Dec 06 11:40:47 crc kubenswrapper[4678]: I1206 11:40:47.614305 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033"} err="failed to get container status \"0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033\": rpc error: code = NotFound desc = could not find container \"0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033\": container with ID starting with 0ae5fb0dc00a6e302af9b5bce849efaf6ee6f99032cfc863773c1ae1ff263033 not found: ID does not exist" Dec 06 11:40:49 crc kubenswrapper[4678]: I1206 11:40:49.500067 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" path="/var/lib/kubelet/pods/af22b429-f8e3-45cf-af1f-b37660da92c4/volumes" Dec 06 11:40:56 crc kubenswrapper[4678]: E1206 11:40:56.432094 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf22b429_f8e3_45cf_af1f_b37660da92c4.slice/crio-5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:40:56 crc kubenswrapper[4678]: I1206 11:40:56.476169 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:40:56 crc kubenswrapper[4678]: E1206 11:40:56.476552 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:41:07 crc kubenswrapper[4678]: E1206 11:41:06.648866 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf22b429_f8e3_45cf_af1f_b37660da92c4.slice/crio-5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:41:10 crc kubenswrapper[4678]: I1206 11:41:10.478162 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:41:10 crc kubenswrapper[4678]: E1206 11:41:10.478987 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:41:16 crc kubenswrapper[4678]: E1206 11:41:16.887797 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf22b429_f8e3_45cf_af1f_b37660da92c4.slice/crio-5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:41:21 crc kubenswrapper[4678]: I1206 11:41:21.475954 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:41:21 crc kubenswrapper[4678]: E1206 11:41:21.476694 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.426037 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xbn84"] Dec 06 11:41:26 crc kubenswrapper[4678]: E1206 11:41:26.427179 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="registry-server" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427213 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="registry-server" Dec 06 11:41:26 crc kubenswrapper[4678]: E1206 11:41:26.427244 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="registry-server" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427261 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="registry-server" Dec 06 11:41:26 crc kubenswrapper[4678]: E1206 11:41:26.427285 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="extract-utilities" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427305 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="extract-utilities" Dec 06 11:41:26 crc kubenswrapper[4678]: E1206 11:41:26.427333 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="extract-utilities" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427350 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="extract-utilities" Dec 06 11:41:26 crc kubenswrapper[4678]: E1206 11:41:26.427374 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="extract-content" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427390 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="extract-content" Dec 06 11:41:26 crc kubenswrapper[4678]: E1206 11:41:26.427425 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="extract-content" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427442 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="extract-content" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427951 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="af22b429-f8e3-45cf-af1f-b37660da92c4" containerName="registry-server" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.427987 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a70063b-28fa-4ffc-88c2-29d790e853e6" containerName="registry-server" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.430071 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.449720 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xbn84"] Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.551587 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-utilities\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.551630 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqtgf\" (UniqueName: \"kubernetes.io/projected/d593b1f1-f2b1-4691-9e58-0c277c44c448-kube-api-access-rqtgf\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.551668 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-catalog-content\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.653304 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-utilities\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.653366 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqtgf\" (UniqueName: \"kubernetes.io/projected/d593b1f1-f2b1-4691-9e58-0c277c44c448-kube-api-access-rqtgf\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.653414 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-catalog-content\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.655340 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-catalog-content\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.655538 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-utilities\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.676661 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqtgf\" (UniqueName: \"kubernetes.io/projected/d593b1f1-f2b1-4691-9e58-0c277c44c448-kube-api-access-rqtgf\") pod \"redhat-operators-xbn84\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:26 crc kubenswrapper[4678]: I1206 11:41:26.747338 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:27 crc kubenswrapper[4678]: E1206 11:41:27.137749 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf22b429_f8e3_45cf_af1f_b37660da92c4.slice/crio-5c137e37a9309a155e4b1ec279682c116d5cdc64ebe56f2b61d3bf29b667349e.scope\": RecentStats: unable to find data in memory cache]" Dec 06 11:41:27 crc kubenswrapper[4678]: I1206 11:41:27.311397 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xbn84"] Dec 06 11:41:27 crc kubenswrapper[4678]: I1206 11:41:27.619540 4678 generic.go:334] "Generic (PLEG): container finished" podID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerID="55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57" exitCode=0 Dec 06 11:41:27 crc kubenswrapper[4678]: I1206 11:41:27.619580 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbn84" event={"ID":"d593b1f1-f2b1-4691-9e58-0c277c44c448","Type":"ContainerDied","Data":"55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57"} Dec 06 11:41:27 crc kubenswrapper[4678]: I1206 11:41:27.619603 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbn84" event={"ID":"d593b1f1-f2b1-4691-9e58-0c277c44c448","Type":"ContainerStarted","Data":"5e94528ce6ef30fa8922bbe9d53b0dea0007436ffd0bca1839f90e312332809c"} Dec 06 11:41:28 crc kubenswrapper[4678]: I1206 11:41:28.628420 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbn84" event={"ID":"d593b1f1-f2b1-4691-9e58-0c277c44c448","Type":"ContainerStarted","Data":"bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396"} Dec 06 11:41:32 crc kubenswrapper[4678]: I1206 11:41:32.667103 4678 generic.go:334] "Generic (PLEG): container finished" podID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerID="bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396" exitCode=0 Dec 06 11:41:32 crc kubenswrapper[4678]: I1206 11:41:32.667173 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbn84" event={"ID":"d593b1f1-f2b1-4691-9e58-0c277c44c448","Type":"ContainerDied","Data":"bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396"} Dec 06 11:41:33 crc kubenswrapper[4678]: I1206 11:41:33.681284 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbn84" event={"ID":"d593b1f1-f2b1-4691-9e58-0c277c44c448","Type":"ContainerStarted","Data":"5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c"} Dec 06 11:41:33 crc kubenswrapper[4678]: I1206 11:41:33.703319 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xbn84" podStartSLOduration=2.2663461590000002 podStartE2EDuration="7.703293866s" podCreationTimestamp="2025-12-06 11:41:26 +0000 UTC" firstStartedPulling="2025-12-06 11:41:27.621687117 +0000 UTC m=+3892.465118556" lastFinishedPulling="2025-12-06 11:41:33.058634824 +0000 UTC m=+3897.902066263" observedRunningTime="2025-12-06 11:41:33.698334969 +0000 UTC m=+3898.541766448" watchObservedRunningTime="2025-12-06 11:41:33.703293866 +0000 UTC m=+3898.546725345" Dec 06 11:41:35 crc kubenswrapper[4678]: I1206 11:41:35.491021 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:41:35 crc kubenswrapper[4678]: E1206 11:41:35.496039 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:41:36 crc kubenswrapper[4678]: I1206 11:41:36.747539 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:36 crc kubenswrapper[4678]: I1206 11:41:36.747883 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:37 crc kubenswrapper[4678]: I1206 11:41:37.812793 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xbn84" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="registry-server" probeResult="failure" output=< Dec 06 11:41:37 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:41:37 crc kubenswrapper[4678]: > Dec 06 11:41:46 crc kubenswrapper[4678]: I1206 11:41:46.844570 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:46 crc kubenswrapper[4678]: I1206 11:41:46.950312 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:47 crc kubenswrapper[4678]: I1206 11:41:47.112740 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xbn84"] Dec 06 11:41:47 crc kubenswrapper[4678]: I1206 11:41:47.791503 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="747aae14-74af-41d1-af19-610b0fa68be1" containerName="galera" probeResult="failure" output="command timed out" Dec 06 11:41:48 crc kubenswrapper[4678]: I1206 11:41:48.476587 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:41:48 crc kubenswrapper[4678]: E1206 11:41:48.476903 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:41:48 crc kubenswrapper[4678]: I1206 11:41:48.769236 4678 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1cb14647-bb80-422d-8f59-985a70bf14c3" containerName="galera" probeResult="failure" output="command timed out" Dec 06 11:41:48 crc kubenswrapper[4678]: I1206 11:41:48.770288 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1cb14647-bb80-422d-8f59-985a70bf14c3" containerName="galera" probeResult="failure" output="command timed out" Dec 06 11:41:48 crc kubenswrapper[4678]: I1206 11:41:48.830313 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xbn84" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="registry-server" containerID="cri-o://5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c" gracePeriod=2 Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.772349 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.820489 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqtgf\" (UniqueName: \"kubernetes.io/projected/d593b1f1-f2b1-4691-9e58-0c277c44c448-kube-api-access-rqtgf\") pod \"d593b1f1-f2b1-4691-9e58-0c277c44c448\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.820638 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-catalog-content\") pod \"d593b1f1-f2b1-4691-9e58-0c277c44c448\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.820786 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-utilities\") pod \"d593b1f1-f2b1-4691-9e58-0c277c44c448\" (UID: \"d593b1f1-f2b1-4691-9e58-0c277c44c448\") " Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.821966 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-utilities" (OuterVolumeSpecName: "utilities") pod "d593b1f1-f2b1-4691-9e58-0c277c44c448" (UID: "d593b1f1-f2b1-4691-9e58-0c277c44c448"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.832808 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d593b1f1-f2b1-4691-9e58-0c277c44c448-kube-api-access-rqtgf" (OuterVolumeSpecName: "kube-api-access-rqtgf") pod "d593b1f1-f2b1-4691-9e58-0c277c44c448" (UID: "d593b1f1-f2b1-4691-9e58-0c277c44c448"). InnerVolumeSpecName "kube-api-access-rqtgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.847108 4678 generic.go:334] "Generic (PLEG): container finished" podID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerID="5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c" exitCode=0 Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.847161 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbn84" event={"ID":"d593b1f1-f2b1-4691-9e58-0c277c44c448","Type":"ContainerDied","Data":"5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c"} Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.847223 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xbn84" event={"ID":"d593b1f1-f2b1-4691-9e58-0c277c44c448","Type":"ContainerDied","Data":"5e94528ce6ef30fa8922bbe9d53b0dea0007436ffd0bca1839f90e312332809c"} Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.847684 4678 scope.go:117] "RemoveContainer" containerID="5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.847903 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xbn84" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.915238 4678 scope.go:117] "RemoveContainer" containerID="bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.924039 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.925168 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqtgf\" (UniqueName: \"kubernetes.io/projected/d593b1f1-f2b1-4691-9e58-0c277c44c448-kube-api-access-rqtgf\") on node \"crc\" DevicePath \"\"" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.963742 4678 scope.go:117] "RemoveContainer" containerID="55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.978000 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d593b1f1-f2b1-4691-9e58-0c277c44c448" (UID: "d593b1f1-f2b1-4691-9e58-0c277c44c448"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.996540 4678 scope.go:117] "RemoveContainer" containerID="5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c" Dec 06 11:41:49 crc kubenswrapper[4678]: E1206 11:41:49.998525 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c\": container with ID starting with 5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c not found: ID does not exist" containerID="5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.998640 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c"} err="failed to get container status \"5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c\": rpc error: code = NotFound desc = could not find container \"5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c\": container with ID starting with 5f2231ecf7937bff7ea3d9f60327329a60a9a68539952c4f72a048e638e0224c not found: ID does not exist" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.998725 4678 scope.go:117] "RemoveContainer" containerID="bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396" Dec 06 11:41:49 crc kubenswrapper[4678]: E1206 11:41:49.999714 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396\": container with ID starting with bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396 not found: ID does not exist" containerID="bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.999820 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396"} err="failed to get container status \"bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396\": rpc error: code = NotFound desc = could not find container \"bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396\": container with ID starting with bd64ba4730198d21a5822c7ca32b5c450ab6d8dc6d1857c61efec0d8b81fb396 not found: ID does not exist" Dec 06 11:41:49 crc kubenswrapper[4678]: I1206 11:41:49.999896 4678 scope.go:117] "RemoveContainer" containerID="55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57" Dec 06 11:41:50 crc kubenswrapper[4678]: E1206 11:41:50.000200 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57\": container with ID starting with 55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57 not found: ID does not exist" containerID="55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57" Dec 06 11:41:50 crc kubenswrapper[4678]: I1206 11:41:50.000242 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57"} err="failed to get container status \"55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57\": rpc error: code = NotFound desc = could not find container \"55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57\": container with ID starting with 55dff28d1160a83441297f6e2720235eeee232f91cddf561bd04cdefafd1af57 not found: ID does not exist" Dec 06 11:41:50 crc kubenswrapper[4678]: I1206 11:41:50.027697 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d593b1f1-f2b1-4691-9e58-0c277c44c448-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:41:50 crc kubenswrapper[4678]: I1206 11:41:50.183814 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xbn84"] Dec 06 11:41:50 crc kubenswrapper[4678]: I1206 11:41:50.192943 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xbn84"] Dec 06 11:41:51 crc kubenswrapper[4678]: I1206 11:41:51.488325 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" path="/var/lib/kubelet/pods/d593b1f1-f2b1-4691-9e58-0c277c44c448/volumes" Dec 06 11:42:01 crc kubenswrapper[4678]: I1206 11:42:01.475909 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:42:01 crc kubenswrapper[4678]: E1206 11:42:01.476688 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:42:13 crc kubenswrapper[4678]: I1206 11:42:13.475810 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:42:13 crc kubenswrapper[4678]: E1206 11:42:13.476672 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:42:26 crc kubenswrapper[4678]: I1206 11:42:26.476062 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:42:26 crc kubenswrapper[4678]: E1206 11:42:26.476816 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:42:37 crc kubenswrapper[4678]: I1206 11:42:37.476190 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:42:38 crc kubenswrapper[4678]: I1206 11:42:38.274267 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"841eeeda6a566780d59bc42c1531aafed9b91726537c5ad525315b37460eacdd"} Dec 06 11:44:59 crc kubenswrapper[4678]: I1206 11:44:59.505969 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:44:59 crc kubenswrapper[4678]: I1206 11:44:59.506420 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.184304 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn"] Dec 06 11:45:00 crc kubenswrapper[4678]: E1206 11:45:00.184845 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="registry-server" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.184865 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="registry-server" Dec 06 11:45:00 crc kubenswrapper[4678]: E1206 11:45:00.184886 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="extract-content" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.184895 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="extract-content" Dec 06 11:45:00 crc kubenswrapper[4678]: E1206 11:45:00.184909 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="extract-utilities" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.184918 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="extract-utilities" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.185180 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="d593b1f1-f2b1-4691-9e58-0c277c44c448" containerName="registry-server" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.187002 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.190936 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.191273 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.204724 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn"] Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.247029 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv52d\" (UniqueName: \"kubernetes.io/projected/a8168df3-62dd-460d-b33a-f27c4f0b44b0-kube-api-access-wv52d\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.247088 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8168df3-62dd-460d-b33a-f27c4f0b44b0-secret-volume\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.247121 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8168df3-62dd-460d-b33a-f27c4f0b44b0-config-volume\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.348631 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv52d\" (UniqueName: \"kubernetes.io/projected/a8168df3-62dd-460d-b33a-f27c4f0b44b0-kube-api-access-wv52d\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.348717 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8168df3-62dd-460d-b33a-f27c4f0b44b0-secret-volume\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.348755 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8168df3-62dd-460d-b33a-f27c4f0b44b0-config-volume\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.349636 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8168df3-62dd-460d-b33a-f27c4f0b44b0-config-volume\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.354882 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8168df3-62dd-460d-b33a-f27c4f0b44b0-secret-volume\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.366520 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv52d\" (UniqueName: \"kubernetes.io/projected/a8168df3-62dd-460d-b33a-f27c4f0b44b0-kube-api-access-wv52d\") pod \"collect-profiles-29417025-njhmn\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:00 crc kubenswrapper[4678]: I1206 11:45:00.538277 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:01 crc kubenswrapper[4678]: I1206 11:45:01.009860 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn"] Dec 06 11:45:01 crc kubenswrapper[4678]: I1206 11:45:01.650128 4678 generic.go:334] "Generic (PLEG): container finished" podID="a8168df3-62dd-460d-b33a-f27c4f0b44b0" containerID="ed6ad98f07d2febc5aeece60129eb2a552391ae916d31f03aac0eb11011eaf9e" exitCode=0 Dec 06 11:45:01 crc kubenswrapper[4678]: I1206 11:45:01.650187 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" event={"ID":"a8168df3-62dd-460d-b33a-f27c4f0b44b0","Type":"ContainerDied","Data":"ed6ad98f07d2febc5aeece60129eb2a552391ae916d31f03aac0eb11011eaf9e"} Dec 06 11:45:01 crc kubenswrapper[4678]: I1206 11:45:01.650450 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" event={"ID":"a8168df3-62dd-460d-b33a-f27c4f0b44b0","Type":"ContainerStarted","Data":"f3f5b9e68ddf08e549e828ebb1f74f5b6bff8e6f7af9cdf32cd7caa75186a87a"} Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.137942 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.205081 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8168df3-62dd-460d-b33a-f27c4f0b44b0-config-volume\") pod \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.205421 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8168df3-62dd-460d-b33a-f27c4f0b44b0-secret-volume\") pod \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.205684 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv52d\" (UniqueName: \"kubernetes.io/projected/a8168df3-62dd-460d-b33a-f27c4f0b44b0-kube-api-access-wv52d\") pod \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\" (UID: \"a8168df3-62dd-460d-b33a-f27c4f0b44b0\") " Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.208692 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8168df3-62dd-460d-b33a-f27c4f0b44b0-config-volume" (OuterVolumeSpecName: "config-volume") pod "a8168df3-62dd-460d-b33a-f27c4f0b44b0" (UID: "a8168df3-62dd-460d-b33a-f27c4f0b44b0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.211525 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8168df3-62dd-460d-b33a-f27c4f0b44b0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a8168df3-62dd-460d-b33a-f27c4f0b44b0" (UID: "a8168df3-62dd-460d-b33a-f27c4f0b44b0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.212278 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8168df3-62dd-460d-b33a-f27c4f0b44b0-kube-api-access-wv52d" (OuterVolumeSpecName: "kube-api-access-wv52d") pod "a8168df3-62dd-460d-b33a-f27c4f0b44b0" (UID: "a8168df3-62dd-460d-b33a-f27c4f0b44b0"). InnerVolumeSpecName "kube-api-access-wv52d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.308073 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv52d\" (UniqueName: \"kubernetes.io/projected/a8168df3-62dd-460d-b33a-f27c4f0b44b0-kube-api-access-wv52d\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.308334 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8168df3-62dd-460d-b33a-f27c4f0b44b0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.308396 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8168df3-62dd-460d-b33a-f27c4f0b44b0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.674747 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" event={"ID":"a8168df3-62dd-460d-b33a-f27c4f0b44b0","Type":"ContainerDied","Data":"f3f5b9e68ddf08e549e828ebb1f74f5b6bff8e6f7af9cdf32cd7caa75186a87a"} Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.674794 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3f5b9e68ddf08e549e828ebb1f74f5b6bff8e6f7af9cdf32cd7caa75186a87a" Dec 06 11:45:03 crc kubenswrapper[4678]: I1206 11:45:03.674801 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417025-njhmn" Dec 06 11:45:04 crc kubenswrapper[4678]: I1206 11:45:04.250433 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2"] Dec 06 11:45:04 crc kubenswrapper[4678]: I1206 11:45:04.273067 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416980-x7ft2"] Dec 06 11:45:05 crc kubenswrapper[4678]: I1206 11:45:05.489975 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="096244e2-6040-4775-a2c3-29da87ed9817" path="/var/lib/kubelet/pods/096244e2-6040-4775-a2c3-29da87ed9817/volumes" Dec 06 11:45:09 crc kubenswrapper[4678]: I1206 11:45:09.738599 4678 generic.go:334] "Generic (PLEG): container finished" podID="5b32697e-d8b7-4852-8afe-a6f19920f210" containerID="4f70958e4fa5a118a4e6f302724c076b674246de73a65eb898bc894a669b3ff2" exitCode=0 Dec 06 11:45:09 crc kubenswrapper[4678]: I1206 11:45:09.738732 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5b32697e-d8b7-4852-8afe-a6f19920f210","Type":"ContainerDied","Data":"4f70958e4fa5a118a4e6f302724c076b674246de73a65eb898bc894a669b3ff2"} Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.166471 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.170980 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.171055 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-workdir\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.171108 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44qxb\" (UniqueName: \"kubernetes.io/projected/5b32697e-d8b7-4852-8afe-a6f19920f210-kube-api-access-44qxb\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.171162 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-config-data\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.171222 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.171267 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ssh-key\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.171319 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-temporary\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.172267 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.173286 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-config-data" (OuterVolumeSpecName: "config-data") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.177125 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.179037 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.179803 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b32697e-d8b7-4852-8afe-a6f19920f210-kube-api-access-44qxb" (OuterVolumeSpecName: "kube-api-access-44qxb") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "kube-api-access-44qxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.216007 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.231129 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.272587 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ca-certs\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.272642 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config-secret\") pod \"5b32697e-d8b7-4852-8afe-a6f19920f210\" (UID: \"5b32697e-d8b7-4852-8afe-a6f19920f210\") " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.273319 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.273340 4678 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.273357 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44qxb\" (UniqueName: \"kubernetes.io/projected/5b32697e-d8b7-4852-8afe-a6f19920f210-kube-api-access-44qxb\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.273370 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5b32697e-d8b7-4852-8afe-a6f19920f210-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.273395 4678 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.273407 4678 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.273419 4678 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/5b32697e-d8b7-4852-8afe-a6f19920f210-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.295120 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.295445 4678 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.309611 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "5b32697e-d8b7-4852-8afe-a6f19920f210" (UID: "5b32697e-d8b7-4852-8afe-a6f19920f210"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.375209 4678 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.375258 4678 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.375280 4678 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/5b32697e-d8b7-4852-8afe-a6f19920f210-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.763774 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"5b32697e-d8b7-4852-8afe-a6f19920f210","Type":"ContainerDied","Data":"4628b8f635c1686c6e116c742744f1d081e792b40af640372ec7e9fd6947253d"} Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.763842 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4628b8f635c1686c6e116c742744f1d081e792b40af640372ec7e9fd6947253d" Dec 06 11:45:11 crc kubenswrapper[4678]: I1206 11:45:11.763844 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.572960 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 11:45:16 crc kubenswrapper[4678]: E1206 11:45:16.574115 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8168df3-62dd-460d-b33a-f27c4f0b44b0" containerName="collect-profiles" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.574135 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8168df3-62dd-460d-b33a-f27c4f0b44b0" containerName="collect-profiles" Dec 06 11:45:16 crc kubenswrapper[4678]: E1206 11:45:16.574161 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b32697e-d8b7-4852-8afe-a6f19920f210" containerName="tempest-tests-tempest-tests-runner" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.574174 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b32697e-d8b7-4852-8afe-a6f19920f210" containerName="tempest-tests-tempest-tests-runner" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.574484 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8168df3-62dd-460d-b33a-f27c4f0b44b0" containerName="collect-profiles" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.574654 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b32697e-d8b7-4852-8afe-a6f19920f210" containerName="tempest-tests-tempest-tests-runner" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.575890 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.579996 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-d84d5" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.588053 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.677403 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4zdb\" (UniqueName: \"kubernetes.io/projected/3b09e665-d6c6-4ac4-831b-210905b8fa20-kube-api-access-z4zdb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3b09e665-d6c6-4ac4-831b-210905b8fa20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.677452 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3b09e665-d6c6-4ac4-831b-210905b8fa20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.779694 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4zdb\" (UniqueName: \"kubernetes.io/projected/3b09e665-d6c6-4ac4-831b-210905b8fa20-kube-api-access-z4zdb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3b09e665-d6c6-4ac4-831b-210905b8fa20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.779742 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3b09e665-d6c6-4ac4-831b-210905b8fa20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.780532 4678 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3b09e665-d6c6-4ac4-831b-210905b8fa20\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.812248 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3b09e665-d6c6-4ac4-831b-210905b8fa20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.868205 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4zdb\" (UniqueName: \"kubernetes.io/projected/3b09e665-d6c6-4ac4-831b-210905b8fa20-kube-api-access-z4zdb\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"3b09e665-d6c6-4ac4-831b-210905b8fa20\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:16 crc kubenswrapper[4678]: I1206 11:45:16.912961 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 06 11:45:17 crc kubenswrapper[4678]: I1206 11:45:17.343443 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 06 11:45:17 crc kubenswrapper[4678]: I1206 11:45:17.860193 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3b09e665-d6c6-4ac4-831b-210905b8fa20","Type":"ContainerStarted","Data":"9ae0ea1abba007e7817e4f27f6a30e23cc275cf0454fe49e8cad88f6110e9ace"} Dec 06 11:45:18 crc kubenswrapper[4678]: I1206 11:45:18.870378 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"3b09e665-d6c6-4ac4-831b-210905b8fa20","Type":"ContainerStarted","Data":"3cc5ca3c276544c596567e1a588413a8528bc221819b02c628be8080a0a8f161"} Dec 06 11:45:18 crc kubenswrapper[4678]: I1206 11:45:18.888580 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.110208945 podStartE2EDuration="2.888558429s" podCreationTimestamp="2025-12-06 11:45:16 +0000 UTC" firstStartedPulling="2025-12-06 11:45:17.346999451 +0000 UTC m=+4122.190430890" lastFinishedPulling="2025-12-06 11:45:18.125348925 +0000 UTC m=+4122.968780374" observedRunningTime="2025-12-06 11:45:18.884071831 +0000 UTC m=+4123.727503280" watchObservedRunningTime="2025-12-06 11:45:18.888558429 +0000 UTC m=+4123.731989868" Dec 06 11:45:29 crc kubenswrapper[4678]: I1206 11:45:29.505908 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:45:29 crc kubenswrapper[4678]: I1206 11:45:29.506623 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:45:41 crc kubenswrapper[4678]: I1206 11:45:41.989731 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ngf8/must-gather-xqtv2"] Dec 06 11:45:41 crc kubenswrapper[4678]: I1206 11:45:41.991693 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.000413 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8ngf8"/"openshift-service-ca.crt" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.004581 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-8ngf8"/"default-dockercfg-szhgm" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.005113 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-8ngf8"/"kube-root-ca.crt" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.012703 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8ngf8/must-gather-xqtv2"] Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.112816 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af614e30-ef67-4ebc-bf1c-7203ba99b625-must-gather-output\") pod \"must-gather-xqtv2\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.113235 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dm686\" (UniqueName: \"kubernetes.io/projected/af614e30-ef67-4ebc-bf1c-7203ba99b625-kube-api-access-dm686\") pod \"must-gather-xqtv2\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.214666 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af614e30-ef67-4ebc-bf1c-7203ba99b625-must-gather-output\") pod \"must-gather-xqtv2\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.215064 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dm686\" (UniqueName: \"kubernetes.io/projected/af614e30-ef67-4ebc-bf1c-7203ba99b625-kube-api-access-dm686\") pod \"must-gather-xqtv2\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.215135 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af614e30-ef67-4ebc-bf1c-7203ba99b625-must-gather-output\") pod \"must-gather-xqtv2\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.235224 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dm686\" (UniqueName: \"kubernetes.io/projected/af614e30-ef67-4ebc-bf1c-7203ba99b625-kube-api-access-dm686\") pod \"must-gather-xqtv2\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.316462 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.783439 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-8ngf8/must-gather-xqtv2"] Dec 06 11:45:42 crc kubenswrapper[4678]: I1206 11:45:42.797761 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:45:43 crc kubenswrapper[4678]: I1206 11:45:43.149092 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" event={"ID":"af614e30-ef67-4ebc-bf1c-7203ba99b625","Type":"ContainerStarted","Data":"c9c5887eb3d52ec5d80629f60025a26100565de3fb393d6d046c500f7ef05045"} Dec 06 11:45:48 crc kubenswrapper[4678]: I1206 11:45:48.190453 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" event={"ID":"af614e30-ef67-4ebc-bf1c-7203ba99b625","Type":"ContainerStarted","Data":"b453a8c10e76d33366016dca59cb9ab158f42dc8394a0aa13ece77400883b35c"} Dec 06 11:45:48 crc kubenswrapper[4678]: I1206 11:45:48.190877 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" event={"ID":"af614e30-ef67-4ebc-bf1c-7203ba99b625","Type":"ContainerStarted","Data":"c796f2dad127e1fea701eecab88d92e927db3b4d9d4b36c2502ab66c32fa4602"} Dec 06 11:45:48 crc kubenswrapper[4678]: I1206 11:45:48.213148 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" podStartSLOduration=2.708525368 podStartE2EDuration="7.213130611s" podCreationTimestamp="2025-12-06 11:45:41 +0000 UTC" firstStartedPulling="2025-12-06 11:45:42.797393885 +0000 UTC m=+4147.640825314" lastFinishedPulling="2025-12-06 11:45:47.301999078 +0000 UTC m=+4152.145430557" observedRunningTime="2025-12-06 11:45:48.212516793 +0000 UTC m=+4153.055948232" watchObservedRunningTime="2025-12-06 11:45:48.213130611 +0000 UTC m=+4153.056562050" Dec 06 11:45:48 crc kubenswrapper[4678]: I1206 11:45:48.560680 4678 scope.go:117] "RemoveContainer" containerID="b8a5cad4925949cd33b025ecefb6d9e5e2ea129563692cf6c0f2269fd62fc4b9" Dec 06 11:45:51 crc kubenswrapper[4678]: I1206 11:45:51.899821 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-qf99n"] Dec 06 11:45:51 crc kubenswrapper[4678]: I1206 11:45:51.901523 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: I1206 11:45:52.023818 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwvnh\" (UniqueName: \"kubernetes.io/projected/0cfd1018-582a-459e-940a-322445698c4f-kube-api-access-fwvnh\") pod \"crc-debug-qf99n\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: I1206 11:45:52.023916 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfd1018-582a-459e-940a-322445698c4f-host\") pod \"crc-debug-qf99n\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: I1206 11:45:52.125712 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfd1018-582a-459e-940a-322445698c4f-host\") pod \"crc-debug-qf99n\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: I1206 11:45:52.125868 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwvnh\" (UniqueName: \"kubernetes.io/projected/0cfd1018-582a-459e-940a-322445698c4f-kube-api-access-fwvnh\") pod \"crc-debug-qf99n\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: I1206 11:45:52.125863 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfd1018-582a-459e-940a-322445698c4f-host\") pod \"crc-debug-qf99n\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: I1206 11:45:52.164571 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwvnh\" (UniqueName: \"kubernetes.io/projected/0cfd1018-582a-459e-940a-322445698c4f-kube-api-access-fwvnh\") pod \"crc-debug-qf99n\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: I1206 11:45:52.220224 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:45:52 crc kubenswrapper[4678]: W1206 11:45:52.263614 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cfd1018_582a_459e_940a_322445698c4f.slice/crio-d7ec8945d03256e9fc708873e2e98475df2efff6216a1415900e2fb1c99ff9f6 WatchSource:0}: Error finding container d7ec8945d03256e9fc708873e2e98475df2efff6216a1415900e2fb1c99ff9f6: Status 404 returned error can't find the container with id d7ec8945d03256e9fc708873e2e98475df2efff6216a1415900e2fb1c99ff9f6 Dec 06 11:45:53 crc kubenswrapper[4678]: I1206 11:45:53.246835 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" event={"ID":"0cfd1018-582a-459e-940a-322445698c4f","Type":"ContainerStarted","Data":"d7ec8945d03256e9fc708873e2e98475df2efff6216a1415900e2fb1c99ff9f6"} Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.392115 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ssxl2"] Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.394747 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.414150 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ssxl2"] Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.455516 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-catalog-content\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.456442 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-utilities\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.456606 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6s4x\" (UniqueName: \"kubernetes.io/projected/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-kube-api-access-t6s4x\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.558433 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-utilities\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.558720 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6s4x\" (UniqueName: \"kubernetes.io/projected/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-kube-api-access-t6s4x\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.558900 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-catalog-content\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.559632 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-utilities\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.559701 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-catalog-content\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.671453 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6s4x\" (UniqueName: \"kubernetes.io/projected/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-kube-api-access-t6s4x\") pod \"certified-operators-ssxl2\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:57 crc kubenswrapper[4678]: I1206 11:45:57.735672 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:45:58 crc kubenswrapper[4678]: I1206 11:45:58.330226 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ssxl2"] Dec 06 11:45:58 crc kubenswrapper[4678]: W1206 11:45:58.357836 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6d5da93_8ebf_45ff_875c_aff6ad4ee14a.slice/crio-265da63be98d045a2f6a8d77996fb49315fff8509846cdc34b78472b66dddbf0 WatchSource:0}: Error finding container 265da63be98d045a2f6a8d77996fb49315fff8509846cdc34b78472b66dddbf0: Status 404 returned error can't find the container with id 265da63be98d045a2f6a8d77996fb49315fff8509846cdc34b78472b66dddbf0 Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.304187 4678 generic.go:334] "Generic (PLEG): container finished" podID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerID="4f3e0435b5d82b1deb30ae035af4c71e6b50a0c866b953757f829775a3c2772f" exitCode=0 Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.304390 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssxl2" event={"ID":"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a","Type":"ContainerDied","Data":"4f3e0435b5d82b1deb30ae035af4c71e6b50a0c866b953757f829775a3c2772f"} Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.304694 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssxl2" event={"ID":"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a","Type":"ContainerStarted","Data":"265da63be98d045a2f6a8d77996fb49315fff8509846cdc34b78472b66dddbf0"} Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.506073 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.506137 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.506175 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.506621 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"841eeeda6a566780d59bc42c1531aafed9b91726537c5ad525315b37460eacdd"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:45:59 crc kubenswrapper[4678]: I1206 11:45:59.506912 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://841eeeda6a566780d59bc42c1531aafed9b91726537c5ad525315b37460eacdd" gracePeriod=600 Dec 06 11:46:00 crc kubenswrapper[4678]: I1206 11:46:00.315874 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="841eeeda6a566780d59bc42c1531aafed9b91726537c5ad525315b37460eacdd" exitCode=0 Dec 06 11:46:00 crc kubenswrapper[4678]: I1206 11:46:00.315964 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"841eeeda6a566780d59bc42c1531aafed9b91726537c5ad525315b37460eacdd"} Dec 06 11:46:00 crc kubenswrapper[4678]: I1206 11:46:00.316298 4678 scope.go:117] "RemoveContainer" containerID="3ce8fa3b397e286656a0f6e8c7726e66d21c8a270ec0791afb0a69d9b6a48eb1" Dec 06 11:46:06 crc kubenswrapper[4678]: I1206 11:46:06.369411 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" event={"ID":"0cfd1018-582a-459e-940a-322445698c4f","Type":"ContainerStarted","Data":"2ae7b1674831dac6a0dd5675b415f1e4e8597d340747778b7ca2e834cae71029"} Dec 06 11:46:06 crc kubenswrapper[4678]: I1206 11:46:06.373562 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628"} Dec 06 11:46:06 crc kubenswrapper[4678]: I1206 11:46:06.375625 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssxl2" event={"ID":"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a","Type":"ContainerStarted","Data":"27e91936bf33627ce404139dbfa6691e445820fe5cdf7dad9a2862e441c71d00"} Dec 06 11:46:06 crc kubenswrapper[4678]: I1206 11:46:06.391807 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" podStartSLOduration=2.270885543 podStartE2EDuration="15.391782264s" podCreationTimestamp="2025-12-06 11:45:51 +0000 UTC" firstStartedPulling="2025-12-06 11:45:52.266556989 +0000 UTC m=+4157.109988428" lastFinishedPulling="2025-12-06 11:46:05.38745371 +0000 UTC m=+4170.230885149" observedRunningTime="2025-12-06 11:46:06.386303534 +0000 UTC m=+4171.229734973" watchObservedRunningTime="2025-12-06 11:46:06.391782264 +0000 UTC m=+4171.235213703" Dec 06 11:46:08 crc kubenswrapper[4678]: I1206 11:46:08.398095 4678 generic.go:334] "Generic (PLEG): container finished" podID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerID="27e91936bf33627ce404139dbfa6691e445820fe5cdf7dad9a2862e441c71d00" exitCode=0 Dec 06 11:46:08 crc kubenswrapper[4678]: I1206 11:46:08.398294 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssxl2" event={"ID":"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a","Type":"ContainerDied","Data":"27e91936bf33627ce404139dbfa6691e445820fe5cdf7dad9a2862e441c71d00"} Dec 06 11:46:09 crc kubenswrapper[4678]: I1206 11:46:09.421024 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssxl2" event={"ID":"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a","Type":"ContainerStarted","Data":"aac116ff92464dbecb1b447113862aca923316d9dae3de79eef655fab045da5c"} Dec 06 11:46:09 crc kubenswrapper[4678]: I1206 11:46:09.442869 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ssxl2" podStartSLOduration=8.91423232 podStartE2EDuration="12.442853959s" podCreationTimestamp="2025-12-06 11:45:57 +0000 UTC" firstStartedPulling="2025-12-06 11:46:05.251512072 +0000 UTC m=+4170.094943521" lastFinishedPulling="2025-12-06 11:46:08.780133711 +0000 UTC m=+4173.623565160" observedRunningTime="2025-12-06 11:46:09.440519957 +0000 UTC m=+4174.283951396" watchObservedRunningTime="2025-12-06 11:46:09.442853959 +0000 UTC m=+4174.286285398" Dec 06 11:46:17 crc kubenswrapper[4678]: I1206 11:46:17.735901 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:46:17 crc kubenswrapper[4678]: I1206 11:46:17.736461 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:46:17 crc kubenswrapper[4678]: I1206 11:46:17.795099 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:46:18 crc kubenswrapper[4678]: I1206 11:46:18.538226 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:46:18 crc kubenswrapper[4678]: I1206 11:46:18.590402 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ssxl2"] Dec 06 11:46:20 crc kubenswrapper[4678]: I1206 11:46:20.502187 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ssxl2" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="registry-server" containerID="cri-o://aac116ff92464dbecb1b447113862aca923316d9dae3de79eef655fab045da5c" gracePeriod=2 Dec 06 11:46:21 crc kubenswrapper[4678]: I1206 11:46:21.515323 4678 generic.go:334] "Generic (PLEG): container finished" podID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerID="aac116ff92464dbecb1b447113862aca923316d9dae3de79eef655fab045da5c" exitCode=0 Dec 06 11:46:21 crc kubenswrapper[4678]: I1206 11:46:21.515510 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssxl2" event={"ID":"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a","Type":"ContainerDied","Data":"aac116ff92464dbecb1b447113862aca923316d9dae3de79eef655fab045da5c"} Dec 06 11:46:21 crc kubenswrapper[4678]: I1206 11:46:21.515984 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ssxl2" event={"ID":"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a","Type":"ContainerDied","Data":"265da63be98d045a2f6a8d77996fb49315fff8509846cdc34b78472b66dddbf0"} Dec 06 11:46:21 crc kubenswrapper[4678]: I1206 11:46:21.516007 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="265da63be98d045a2f6a8d77996fb49315fff8509846cdc34b78472b66dddbf0" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.377283 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.487944 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-catalog-content\") pod \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.488014 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6s4x\" (UniqueName: \"kubernetes.io/projected/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-kube-api-access-t6s4x\") pod \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.488136 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-utilities\") pod \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\" (UID: \"b6d5da93-8ebf-45ff-875c-aff6ad4ee14a\") " Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.489368 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-utilities" (OuterVolumeSpecName: "utilities") pod "b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" (UID: "b6d5da93-8ebf-45ff-875c-aff6ad4ee14a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.524314 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ssxl2" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.537870 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" (UID: "b6d5da93-8ebf-45ff-875c-aff6ad4ee14a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.590745 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.590775 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.674286 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-kube-api-access-t6s4x" (OuterVolumeSpecName: "kube-api-access-t6s4x") pod "b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" (UID: "b6d5da93-8ebf-45ff-875c-aff6ad4ee14a"). InnerVolumeSpecName "kube-api-access-t6s4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.700146 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6s4x\" (UniqueName: \"kubernetes.io/projected/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a-kube-api-access-t6s4x\") on node \"crc\" DevicePath \"\"" Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.891825 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ssxl2"] Dec 06 11:46:22 crc kubenswrapper[4678]: I1206 11:46:22.904579 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ssxl2"] Dec 06 11:46:23 crc kubenswrapper[4678]: I1206 11:46:23.489758 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" path="/var/lib/kubelet/pods/b6d5da93-8ebf-45ff-875c-aff6ad4ee14a/volumes" Dec 06 11:46:58 crc kubenswrapper[4678]: I1206 11:46:58.819935 4678 generic.go:334] "Generic (PLEG): container finished" podID="0cfd1018-582a-459e-940a-322445698c4f" containerID="2ae7b1674831dac6a0dd5675b415f1e4e8597d340747778b7ca2e834cae71029" exitCode=0 Dec 06 11:46:58 crc kubenswrapper[4678]: I1206 11:46:58.820028 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" event={"ID":"0cfd1018-582a-459e-940a-322445698c4f","Type":"ContainerDied","Data":"2ae7b1674831dac6a0dd5675b415f1e4e8597d340747778b7ca2e834cae71029"} Dec 06 11:46:59 crc kubenswrapper[4678]: I1206 11:46:59.943233 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:46:59 crc kubenswrapper[4678]: I1206 11:46:59.975167 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-qf99n"] Dec 06 11:46:59 crc kubenswrapper[4678]: I1206 11:46:59.997815 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-qf99n"] Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.105272 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfd1018-582a-459e-940a-322445698c4f-host\") pod \"0cfd1018-582a-459e-940a-322445698c4f\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.105379 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0cfd1018-582a-459e-940a-322445698c4f-host" (OuterVolumeSpecName: "host") pod "0cfd1018-582a-459e-940a-322445698c4f" (UID: "0cfd1018-582a-459e-940a-322445698c4f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.105412 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwvnh\" (UniqueName: \"kubernetes.io/projected/0cfd1018-582a-459e-940a-322445698c4f-kube-api-access-fwvnh\") pod \"0cfd1018-582a-459e-940a-322445698c4f\" (UID: \"0cfd1018-582a-459e-940a-322445698c4f\") " Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.107346 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0cfd1018-582a-459e-940a-322445698c4f-host\") on node \"crc\" DevicePath \"\"" Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.111610 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cfd1018-582a-459e-940a-322445698c4f-kube-api-access-fwvnh" (OuterVolumeSpecName: "kube-api-access-fwvnh") pod "0cfd1018-582a-459e-940a-322445698c4f" (UID: "0cfd1018-582a-459e-940a-322445698c4f"). InnerVolumeSpecName "kube-api-access-fwvnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.209541 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwvnh\" (UniqueName: \"kubernetes.io/projected/0cfd1018-582a-459e-940a-322445698c4f-kube-api-access-fwvnh\") on node \"crc\" DevicePath \"\"" Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.839584 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7ec8945d03256e9fc708873e2e98475df2efff6216a1415900e2fb1c99ff9f6" Dec 06 11:47:00 crc kubenswrapper[4678]: I1206 11:47:00.839637 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-qf99n" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.244547 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-vstvz"] Dec 06 11:47:01 crc kubenswrapper[4678]: E1206 11:47:01.245181 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cfd1018-582a-459e-940a-322445698c4f" containerName="container-00" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.245195 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cfd1018-582a-459e-940a-322445698c4f" containerName="container-00" Dec 06 11:47:01 crc kubenswrapper[4678]: E1206 11:47:01.245228 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="extract-content" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.245235 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="extract-content" Dec 06 11:47:01 crc kubenswrapper[4678]: E1206 11:47:01.245252 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="registry-server" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.245259 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="registry-server" Dec 06 11:47:01 crc kubenswrapper[4678]: E1206 11:47:01.245271 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="extract-utilities" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.245277 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="extract-utilities" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.245450 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d5da93-8ebf-45ff-875c-aff6ad4ee14a" containerName="registry-server" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.245472 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cfd1018-582a-459e-940a-322445698c4f" containerName="container-00" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.246045 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.329169 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-host\") pod \"crc-debug-vstvz\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.329369 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rctdz\" (UniqueName: \"kubernetes.io/projected/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-kube-api-access-rctdz\") pod \"crc-debug-vstvz\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.431424 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-host\") pod \"crc-debug-vstvz\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.431524 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rctdz\" (UniqueName: \"kubernetes.io/projected/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-kube-api-access-rctdz\") pod \"crc-debug-vstvz\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.431626 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-host\") pod \"crc-debug-vstvz\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.454380 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rctdz\" (UniqueName: \"kubernetes.io/projected/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-kube-api-access-rctdz\") pod \"crc-debug-vstvz\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.487354 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cfd1018-582a-459e-940a-322445698c4f" path="/var/lib/kubelet/pods/0cfd1018-582a-459e-940a-322445698c4f/volumes" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.559754 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.853304 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-vstvz" event={"ID":"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea","Type":"ContainerStarted","Data":"61f56a469e44fd7f45cb7628f32837de6922e4d664bb30f5a8bc2f8cbd4d904d"} Dec 06 11:47:01 crc kubenswrapper[4678]: I1206 11:47:01.853352 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-vstvz" event={"ID":"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea","Type":"ContainerStarted","Data":"d2295034a0a55f286b3c95a14b2422c7e9c2204b4fe43886ecf8e83e97964a7d"} Dec 06 11:47:02 crc kubenswrapper[4678]: I1206 11:47:02.873409 4678 generic.go:334] "Generic (PLEG): container finished" podID="b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea" containerID="61f56a469e44fd7f45cb7628f32837de6922e4d664bb30f5a8bc2f8cbd4d904d" exitCode=0 Dec 06 11:47:02 crc kubenswrapper[4678]: I1206 11:47:02.873519 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-vstvz" event={"ID":"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea","Type":"ContainerDied","Data":"61f56a469e44fd7f45cb7628f32837de6922e4d664bb30f5a8bc2f8cbd4d904d"} Dec 06 11:47:03 crc kubenswrapper[4678]: I1206 11:47:03.988708 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.026890 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-vstvz"] Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.036178 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-vstvz"] Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.072864 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rctdz\" (UniqueName: \"kubernetes.io/projected/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-kube-api-access-rctdz\") pod \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.072950 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-host\") pod \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\" (UID: \"b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea\") " Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.073049 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-host" (OuterVolumeSpecName: "host") pod "b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea" (UID: "b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.073512 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-host\") on node \"crc\" DevicePath \"\"" Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.088109 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-kube-api-access-rctdz" (OuterVolumeSpecName: "kube-api-access-rctdz") pod "b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea" (UID: "b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea"). InnerVolumeSpecName "kube-api-access-rctdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.175165 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rctdz\" (UniqueName: \"kubernetes.io/projected/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea-kube-api-access-rctdz\") on node \"crc\" DevicePath \"\"" Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.889893 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2295034a0a55f286b3c95a14b2422c7e9c2204b4fe43886ecf8e83e97964a7d" Dec 06 11:47:04 crc kubenswrapper[4678]: I1206 11:47:04.889969 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-vstvz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.346689 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-tlfbz"] Dec 06 11:47:05 crc kubenswrapper[4678]: E1206 11:47:05.347117 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea" containerName="container-00" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.347131 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea" containerName="container-00" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.347309 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea" containerName="container-00" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.347976 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.404675 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-647s2\" (UniqueName: \"kubernetes.io/projected/06285c50-7a5c-4a46-9b8d-e42b26f1a088-kube-api-access-647s2\") pod \"crc-debug-tlfbz\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.404946 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06285c50-7a5c-4a46-9b8d-e42b26f1a088-host\") pod \"crc-debug-tlfbz\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.490481 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea" path="/var/lib/kubelet/pods/b29e8d7e-270e-4ac1-8a12-0bc0340ce7ea/volumes" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.507054 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-647s2\" (UniqueName: \"kubernetes.io/projected/06285c50-7a5c-4a46-9b8d-e42b26f1a088-kube-api-access-647s2\") pod \"crc-debug-tlfbz\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.507437 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06285c50-7a5c-4a46-9b8d-e42b26f1a088-host\") pod \"crc-debug-tlfbz\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.507631 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06285c50-7a5c-4a46-9b8d-e42b26f1a088-host\") pod \"crc-debug-tlfbz\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.525347 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-647s2\" (UniqueName: \"kubernetes.io/projected/06285c50-7a5c-4a46-9b8d-e42b26f1a088-kube-api-access-647s2\") pod \"crc-debug-tlfbz\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.667298 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:05 crc kubenswrapper[4678]: W1206 11:47:05.700437 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06285c50_7a5c_4a46_9b8d_e42b26f1a088.slice/crio-2d344f7b81eba32ef3fbf36ca69babeb28b536009db761c9050512ab44073061 WatchSource:0}: Error finding container 2d344f7b81eba32ef3fbf36ca69babeb28b536009db761c9050512ab44073061: Status 404 returned error can't find the container with id 2d344f7b81eba32ef3fbf36ca69babeb28b536009db761c9050512ab44073061 Dec 06 11:47:05 crc kubenswrapper[4678]: I1206 11:47:05.900916 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" event={"ID":"06285c50-7a5c-4a46-9b8d-e42b26f1a088","Type":"ContainerStarted","Data":"2d344f7b81eba32ef3fbf36ca69babeb28b536009db761c9050512ab44073061"} Dec 06 11:47:06 crc kubenswrapper[4678]: I1206 11:47:06.911874 4678 generic.go:334] "Generic (PLEG): container finished" podID="06285c50-7a5c-4a46-9b8d-e42b26f1a088" containerID="9a56bee80fe28da2744296c65d915eaeee904864a2abac3bf58dd9cc3b88b9ef" exitCode=0 Dec 06 11:47:06 crc kubenswrapper[4678]: I1206 11:47:06.911984 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" event={"ID":"06285c50-7a5c-4a46-9b8d-e42b26f1a088","Type":"ContainerDied","Data":"9a56bee80fe28da2744296c65d915eaeee904864a2abac3bf58dd9cc3b88b9ef"} Dec 06 11:47:06 crc kubenswrapper[4678]: I1206 11:47:06.949677 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-tlfbz"] Dec 06 11:47:06 crc kubenswrapper[4678]: I1206 11:47:06.963499 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ngf8/crc-debug-tlfbz"] Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.299740 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.358595 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06285c50-7a5c-4a46-9b8d-e42b26f1a088-host\") pod \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.358728 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06285c50-7a5c-4a46-9b8d-e42b26f1a088-host" (OuterVolumeSpecName: "host") pod "06285c50-7a5c-4a46-9b8d-e42b26f1a088" (UID: "06285c50-7a5c-4a46-9b8d-e42b26f1a088"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.358794 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-647s2\" (UniqueName: \"kubernetes.io/projected/06285c50-7a5c-4a46-9b8d-e42b26f1a088-kube-api-access-647s2\") pod \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\" (UID: \"06285c50-7a5c-4a46-9b8d-e42b26f1a088\") " Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.359208 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/06285c50-7a5c-4a46-9b8d-e42b26f1a088-host\") on node \"crc\" DevicePath \"\"" Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.364478 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06285c50-7a5c-4a46-9b8d-e42b26f1a088-kube-api-access-647s2" (OuterVolumeSpecName: "kube-api-access-647s2") pod "06285c50-7a5c-4a46-9b8d-e42b26f1a088" (UID: "06285c50-7a5c-4a46-9b8d-e42b26f1a088"). InnerVolumeSpecName "kube-api-access-647s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.460669 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-647s2\" (UniqueName: \"kubernetes.io/projected/06285c50-7a5c-4a46-9b8d-e42b26f1a088-kube-api-access-647s2\") on node \"crc\" DevicePath \"\"" Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.943362 4678 scope.go:117] "RemoveContainer" containerID="9a56bee80fe28da2744296c65d915eaeee904864a2abac3bf58dd9cc3b88b9ef" Dec 06 11:47:08 crc kubenswrapper[4678]: I1206 11:47:08.943579 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/crc-debug-tlfbz" Dec 06 11:47:09 crc kubenswrapper[4678]: I1206 11:47:09.487658 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06285c50-7a5c-4a46-9b8d-e42b26f1a088" path="/var/lib/kubelet/pods/06285c50-7a5c-4a46-9b8d-e42b26f1a088/volumes" Dec 06 11:47:29 crc kubenswrapper[4678]: I1206 11:47:29.899921 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7b874f6c6d-jcklm_6e0b0cc3-b79e-4e3a-8459-adba09620269/barbican-api/0.log" Dec 06 11:47:29 crc kubenswrapper[4678]: I1206 11:47:29.962361 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7b874f6c6d-jcklm_6e0b0cc3-b79e-4e3a-8459-adba09620269/barbican-api-log/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.165082 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5674df646-s4z6g_e55dbe47-fb3b-4020-bba4-f8c038e7cc5b/barbican-keystone-listener/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.234394 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5674df646-s4z6g_e55dbe47-fb3b-4020-bba4-f8c038e7cc5b/barbican-keystone-listener-log/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.427596 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-597bfcdc9c-vdzb7_69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957/barbican-worker/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.507825 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-597bfcdc9c-vdzb7_69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957/barbican-worker-log/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.600568 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf_99d806d5-73ca-4ecc-92cb-b3c300cb04b4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.817238 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/ceilometer-central-agent/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.820877 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/ceilometer-notification-agent/0.log" Dec 06 11:47:30 crc kubenswrapper[4678]: I1206 11:47:30.853212 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/proxy-httpd/0.log" Dec 06 11:47:31 crc kubenswrapper[4678]: I1206 11:47:31.065511 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/sg-core/0.log" Dec 06 11:47:31 crc kubenswrapper[4678]: I1206 11:47:31.143803 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b283a032-2cad-40a7-bb76-88a4f3f47472/cinder-api/0.log" Dec 06 11:47:31 crc kubenswrapper[4678]: I1206 11:47:31.212531 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b283a032-2cad-40a7-bb76-88a4f3f47472/cinder-api-log/0.log" Dec 06 11:47:31 crc kubenswrapper[4678]: I1206 11:47:31.399600 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_918bbdd7-35bb-457d-a29f-5c6e4f04b062/probe/0.log" Dec 06 11:47:31 crc kubenswrapper[4678]: I1206 11:47:31.432905 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_918bbdd7-35bb-457d-a29f-5c6e4f04b062/cinder-scheduler/0.log" Dec 06 11:47:31 crc kubenswrapper[4678]: I1206 11:47:31.660474 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj_5fa31310-7e0f-431a-ae13-287fd25a926c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:31 crc kubenswrapper[4678]: I1206 11:47:31.705968 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8_f104b9bb-12fa-4f5c-86ed-2db61119cb55/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:32 crc kubenswrapper[4678]: I1206 11:47:32.158559 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67cb876dc9-5749h_1ab593af-6400-4c78-b798-fd34038f28dd/init/0.log" Dec 06 11:47:32 crc kubenswrapper[4678]: I1206 11:47:32.465353 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-877bb_9ac6052e-a10a-4709-b1be-99bb37d0922d/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:32 crc kubenswrapper[4678]: I1206 11:47:32.499734 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67cb876dc9-5749h_1ab593af-6400-4c78-b798-fd34038f28dd/init/0.log" Dec 06 11:47:32 crc kubenswrapper[4678]: I1206 11:47:32.564915 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67cb876dc9-5749h_1ab593af-6400-4c78-b798-fd34038f28dd/dnsmasq-dns/0.log" Dec 06 11:47:32 crc kubenswrapper[4678]: I1206 11:47:32.759127 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ef3105b5-af55-4aec-99a8-10792740d742/glance-httpd/0.log" Dec 06 11:47:32 crc kubenswrapper[4678]: I1206 11:47:32.812995 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ef3105b5-af55-4aec-99a8-10792740d742/glance-log/0.log" Dec 06 11:47:33 crc kubenswrapper[4678]: I1206 11:47:33.049150 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_965cdf20-a198-4cc8-9a8f-700388dbf4cd/glance-httpd/0.log" Dec 06 11:47:33 crc kubenswrapper[4678]: I1206 11:47:33.058330 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_965cdf20-a198-4cc8-9a8f-700388dbf4cd/glance-log/0.log" Dec 06 11:47:33 crc kubenswrapper[4678]: I1206 11:47:33.209656 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d77bcbbf8-cfc4l_c2ceb969-bd57-4345-840c-4cd8b2b7ca8e/horizon/1.log" Dec 06 11:47:33 crc kubenswrapper[4678]: I1206 11:47:33.390690 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d77bcbbf8-cfc4l_c2ceb969-bd57-4345-840c-4cd8b2b7ca8e/horizon/0.log" Dec 06 11:47:33 crc kubenswrapper[4678]: I1206 11:47:33.533994 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt_a8f9dc0d-5e17-4e1b-9363-2d193c592e9b/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:33 crc kubenswrapper[4678]: I1206 11:47:33.788005 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4l22m_8e6b45b8-d8db-457e-9a08-04ce67b197f6/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:33 crc kubenswrapper[4678]: I1206 11:47:33.864759 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d77bcbbf8-cfc4l_c2ceb969-bd57-4345-840c-4cd8b2b7ca8e/horizon-log/0.log" Dec 06 11:47:34 crc kubenswrapper[4678]: I1206 11:47:34.192252 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29416981-627fr_29f472e5-54ad-470c-a43b-a6fbee7fa24d/keystone-cron/0.log" Dec 06 11:47:34 crc kubenswrapper[4678]: I1206 11:47:34.458978 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3/kube-state-metrics/0.log" Dec 06 11:47:34 crc kubenswrapper[4678]: I1206 11:47:34.709809 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl_ff8f7b93-13e6-49a3-937a-8247a3eaf5c8/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:34 crc kubenswrapper[4678]: I1206 11:47:34.771006 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7f69d54bd9-rd99t_eb641828-16dd-4437-8266-e20a7a27ad1c/keystone-api/0.log" Dec 06 11:47:35 crc kubenswrapper[4678]: I1206 11:47:35.565070 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7567449455-km9cw_0e386ecc-df37-4e08-a9f9-80abb7019c19/neutron-httpd/0.log" Dec 06 11:47:35 crc kubenswrapper[4678]: I1206 11:47:35.754306 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6_1e7abea2-75f9-4bcb-897d-d62fe1e7be3e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:35 crc kubenswrapper[4678]: I1206 11:47:35.770436 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7567449455-km9cw_0e386ecc-df37-4e08-a9f9-80abb7019c19/neutron-api/0.log" Dec 06 11:47:36 crc kubenswrapper[4678]: I1206 11:47:36.722061 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a92133d6-4bca-447f-9b4c-a1e776a4cb40/nova-cell0-conductor-conductor/0.log" Dec 06 11:47:37 crc kubenswrapper[4678]: I1206 11:47:37.066138 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ed1b2724-cbb5-4589-9622-6ff20ab1f180/nova-cell1-conductor-conductor/0.log" Dec 06 11:47:37 crc kubenswrapper[4678]: I1206 11:47:37.406124 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_21d64141-1313-4582-ac36-a0d0a304f0fb/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 11:47:37 crc kubenswrapper[4678]: I1206 11:47:37.439062 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4859acf4-20ed-4af8-a83d-9c6303eec9d4/nova-api-api/0.log" Dec 06 11:47:37 crc kubenswrapper[4678]: I1206 11:47:37.519094 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4859acf4-20ed-4af8-a83d-9c6303eec9d4/nova-api-log/0.log" Dec 06 11:47:37 crc kubenswrapper[4678]: I1206 11:47:37.756340 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-4fnld_ff53c201-513d-4b60-b0ef-a1a59a30bd4f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:37 crc kubenswrapper[4678]: I1206 11:47:37.807095 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_63c277fe-de03-4243-98a2-8d7efaa92c4a/nova-metadata-log/0.log" Dec 06 11:47:38 crc kubenswrapper[4678]: I1206 11:47:38.331738 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1cb14647-bb80-422d-8f59-985a70bf14c3/mysql-bootstrap/0.log" Dec 06 11:47:38 crc kubenswrapper[4678]: I1206 11:47:38.552712 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1cb14647-bb80-422d-8f59-985a70bf14c3/mysql-bootstrap/0.log" Dec 06 11:47:38 crc kubenswrapper[4678]: I1206 11:47:38.665971 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1cb14647-bb80-422d-8f59-985a70bf14c3/galera/0.log" Dec 06 11:47:38 crc kubenswrapper[4678]: I1206 11:47:38.703937 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a2fb742d-13dd-468a-918e-37c53f81b294/nova-scheduler-scheduler/0.log" Dec 06 11:47:38 crc kubenswrapper[4678]: I1206 11:47:38.977314 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_747aae14-74af-41d1-af19-610b0fa68be1/mysql-bootstrap/0.log" Dec 06 11:47:39 crc kubenswrapper[4678]: I1206 11:47:39.245321 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_747aae14-74af-41d1-af19-610b0fa68be1/mysql-bootstrap/0.log" Dec 06 11:47:39 crc kubenswrapper[4678]: I1206 11:47:39.267127 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_747aae14-74af-41d1-af19-610b0fa68be1/galera/0.log" Dec 06 11:47:39 crc kubenswrapper[4678]: I1206 11:47:39.529594 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_66756fbd-c49f-4ea6-8595-7801a7ddecf3/openstackclient/0.log" Dec 06 11:47:39 crc kubenswrapper[4678]: I1206 11:47:39.658326 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kdb5q_c2658168-1355-4595-bf23-570a68c90da1/ovn-controller/0.log" Dec 06 11:47:39 crc kubenswrapper[4678]: I1206 11:47:39.684435 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_63c277fe-de03-4243-98a2-8d7efaa92c4a/nova-metadata-metadata/0.log" Dec 06 11:47:39 crc kubenswrapper[4678]: I1206 11:47:39.827796 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-dgln2_f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69/openstack-network-exporter/0.log" Dec 06 11:47:40 crc kubenswrapper[4678]: I1206 11:47:40.401303 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovsdb-server-init/0.log" Dec 06 11:47:40 crc kubenswrapper[4678]: I1206 11:47:40.795726 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovsdb-server/0.log" Dec 06 11:47:40 crc kubenswrapper[4678]: I1206 11:47:40.810377 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovs-vswitchd/0.log" Dec 06 11:47:40 crc kubenswrapper[4678]: I1206 11:47:40.825387 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovsdb-server-init/0.log" Dec 06 11:47:41 crc kubenswrapper[4678]: I1206 11:47:41.116285 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7xqv9_c79d4229-477e-4f22-90d8-449984bab4cb/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:41 crc kubenswrapper[4678]: I1206 11:47:41.142642 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5a586aa5-fe85-49da-a10e-3e29de7dba4d/openstack-network-exporter/0.log" Dec 06 11:47:41 crc kubenswrapper[4678]: I1206 11:47:41.198462 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5a586aa5-fe85-49da-a10e-3e29de7dba4d/ovn-northd/0.log" Dec 06 11:47:41 crc kubenswrapper[4678]: I1206 11:47:41.444336 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12052ef6-7a13-4ad5-9484-bbad2cabdf5a/openstack-network-exporter/0.log" Dec 06 11:47:41 crc kubenswrapper[4678]: I1206 11:47:41.597073 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12052ef6-7a13-4ad5-9484-bbad2cabdf5a/ovsdbserver-nb/0.log" Dec 06 11:47:41 crc kubenswrapper[4678]: I1206 11:47:41.705703 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3f6c33f7-7b74-437b-a9fd-3790af864af7/ovsdbserver-sb/0.log" Dec 06 11:47:41 crc kubenswrapper[4678]: I1206 11:47:41.777179 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3f6c33f7-7b74-437b-a9fd-3790af864af7/openstack-network-exporter/0.log" Dec 06 11:47:42 crc kubenswrapper[4678]: I1206 11:47:42.147362 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bd747c594-kkr56_e209518e-5689-468e-9b70-9c1e2864eb9b/placement-api/0.log" Dec 06 11:47:42 crc kubenswrapper[4678]: I1206 11:47:42.303617 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bd747c594-kkr56_e209518e-5689-468e-9b70-9c1e2864eb9b/placement-log/0.log" Dec 06 11:47:42 crc kubenswrapper[4678]: I1206 11:47:42.661374 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3/setup-container/0.log" Dec 06 11:47:42 crc kubenswrapper[4678]: I1206 11:47:42.948791 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3/rabbitmq/0.log" Dec 06 11:47:42 crc kubenswrapper[4678]: I1206 11:47:42.975942 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3/setup-container/0.log" Dec 06 11:47:43 crc kubenswrapper[4678]: I1206 11:47:43.019860 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cb4272e8-bb6c-43d4-8539-baeb4f3c14a8/setup-container/0.log" Dec 06 11:47:43 crc kubenswrapper[4678]: I1206 11:47:43.360649 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cb4272e8-bb6c-43d4-8539-baeb4f3c14a8/rabbitmq/0.log" Dec 06 11:47:43 crc kubenswrapper[4678]: I1206 11:47:43.441940 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cb4272e8-bb6c-43d4-8539-baeb4f3c14a8/setup-container/0.log" Dec 06 11:47:43 crc kubenswrapper[4678]: I1206 11:47:43.540181 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz_d6e6e2e2-4745-46f6-a3fa-94f803ac0677/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:44 crc kubenswrapper[4678]: I1206 11:47:44.201838 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-jb8nh_d66737d1-caae-4204-9786-2b19a88f5f82/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:44 crc kubenswrapper[4678]: I1206 11:47:44.263365 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb_29b633cb-d39d-43b9-9556-461cd563c92f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:44 crc kubenswrapper[4678]: I1206 11:47:44.516844 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-7kqwq_fdc4c25f-dc36-4a52-a26d-9dd390c23061/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:44 crc kubenswrapper[4678]: I1206 11:47:44.566243 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9pvlf_06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed/ssh-known-hosts-edpm-deployment/0.log" Dec 06 11:47:44 crc kubenswrapper[4678]: I1206 11:47:44.908808 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f5f494679-7bm9s_b31f2128-db6e-498b-bd38-d0b1c41a4603/proxy-httpd/0.log" Dec 06 11:47:44 crc kubenswrapper[4678]: I1206 11:47:44.931266 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f5f494679-7bm9s_b31f2128-db6e-498b-bd38-d0b1c41a4603/proxy-server/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.103234 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-tkg52_db2561ff-632a-463d-a7af-79d83536812b/swift-ring-rebalance/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.172703 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-auditor/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.215036 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-reaper/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.447531 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-auditor/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.455255 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-replicator/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.548182 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-server/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.553158 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-replicator/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.724764 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-updater/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.799053 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-server/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.835208 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-auditor/0.log" Dec 06 11:47:45 crc kubenswrapper[4678]: I1206 11:47:45.843151 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-expirer/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.005875 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-replicator/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.026073 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-server/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.107531 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-updater/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.183339 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/rsync/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.412283 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/swift-recon-cron/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.438843 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-dj678_1cadf619-4b68-44a3-a65b-422bea0a2ede/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.702723 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_3b09e665-d6c6-4ac4-831b-210905b8fa20/test-operator-logs-container/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.733994 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5b32697e-d8b7-4852-8afe-a6f19920f210/tempest-tests-tempest-tests-runner/0.log" Dec 06 11:47:46 crc kubenswrapper[4678]: I1206 11:47:46.940749 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs_89741e1b-d65e-4e23-982d-716c2e498c25/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:47:58 crc kubenswrapper[4678]: I1206 11:47:58.038718 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a0c958d8-87e6-4b93-be23-58cb84d32dd4/memcached/0.log" Dec 06 11:48:21 crc kubenswrapper[4678]: I1206 11:48:21.695973 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/util/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.289935 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/util/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.324550 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/pull/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.352648 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/pull/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.466994 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/util/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.545808 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/extract/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.549620 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/pull/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.738336 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-th2qw_6669b3a4-9c79-441e-90a2-abceaa89fc58/kube-rbac-proxy/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.831506 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-th2qw_6669b3a4-9c79-441e-90a2-abceaa89fc58/manager/0.log" Dec 06 11:48:22 crc kubenswrapper[4678]: I1206 11:48:22.975387 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zftrw_e48c5af7-e126-49e1-9bc0-eec4bd474080/kube-rbac-proxy/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.047705 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zftrw_e48c5af7-e126-49e1-9bc0-eec4bd474080/manager/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.099967 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-xqpcb_807c6a8a-a12e-4ac6-8d6a-01f54876d4dd/kube-rbac-proxy/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.220759 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-xqpcb_807c6a8a-a12e-4ac6-8d6a-01f54876d4dd/manager/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.324083 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-jpgnt_961960ba-f29c-40e1-81c4-5f8b43456dca/kube-rbac-proxy/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.397685 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-jpgnt_961960ba-f29c-40e1-81c4-5f8b43456dca/manager/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.555722 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-ktmz7_c58c8710-6867-43d2-8fb4-ea96041e2422/kube-rbac-proxy/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.586032 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-ktmz7_c58c8710-6867-43d2-8fb4-ea96041e2422/manager/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.759216 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9kgbd_b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef/kube-rbac-proxy/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.854345 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9kgbd_b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef/manager/0.log" Dec 06 11:48:23 crc kubenswrapper[4678]: I1206 11:48:23.929019 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-9svrt_d272c78c-4c56-4faa-a421-ede41c4b2307/kube-rbac-proxy/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.202700 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-9svrt_d272c78c-4c56-4faa-a421-ede41c4b2307/manager/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.225963 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-ssssw_7bc20ad2-7338-4943-ae21-1ad8a0fd947e/kube-rbac-proxy/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.236375 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-ssssw_7bc20ad2-7338-4943-ae21-1ad8a0fd947e/manager/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.447683 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p5bxv_78fc5203-6af1-49e0-a485-ba9ef0a0b658/kube-rbac-proxy/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.597356 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p5bxv_78fc5203-6af1-49e0-a485-ba9ef0a0b658/manager/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.650958 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-vdspj_9e2b27dd-544c-4c0a-ba0f-6c0b4243017c/kube-rbac-proxy/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.678192 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-vdspj_9e2b27dd-544c-4c0a-ba0f-6c0b4243017c/manager/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.830438 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9b7t6_c7983f9e-7a85-4c2c-b689-7213f06a144a/kube-rbac-proxy/0.log" Dec 06 11:48:24 crc kubenswrapper[4678]: I1206 11:48:24.883248 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9b7t6_c7983f9e-7a85-4c2c-b689-7213f06a144a/manager/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.075097 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-d5h8j_d925b372-ccef-42a2-9b2c-de8539f9d070/kube-rbac-proxy/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.176088 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-d5h8j_d925b372-ccef-42a2-9b2c-de8539f9d070/manager/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.188977 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4kvsf_6424c424-5901-4d42-a077-8388ef0978b8/kube-rbac-proxy/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.404085 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4kvsf_6424c424-5901-4d42-a077-8388ef0978b8/manager/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.470270 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fpj6x_f1b9824d-0c1c-4cef-bb30-3e0f14719c6d/kube-rbac-proxy/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.503723 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fpj6x_f1b9824d-0c1c-4cef-bb30-3e0f14719c6d/manager/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.748219 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fs2699_20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f/kube-rbac-proxy/0.log" Dec 06 11:48:25 crc kubenswrapper[4678]: I1206 11:48:25.760545 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fs2699_20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f/manager/0.log" Dec 06 11:48:26 crc kubenswrapper[4678]: I1206 11:48:26.143034 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-m7422_dfb6011a-3444-4e31-8a84-d2c521f55e77/registry-server/0.log" Dec 06 11:48:26 crc kubenswrapper[4678]: I1206 11:48:26.180425 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bbffdd548-5zgrs_0b336e20-2347-4b86-86b4-22ef7da8983c/operator/0.log" Dec 06 11:48:26 crc kubenswrapper[4678]: I1206 11:48:26.791206 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mrqtn_ed890ee4-e09b-40cd-9bce-f21d8676a2c7/kube-rbac-proxy/0.log" Dec 06 11:48:26 crc kubenswrapper[4678]: I1206 11:48:26.920120 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mrqtn_ed890ee4-e09b-40cd-9bce-f21d8676a2c7/manager/0.log" Dec 06 11:48:26 crc kubenswrapper[4678]: I1206 11:48:26.935055 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5dcc4c7f94-xld2p_67b34972-6c66-47d8-885e-38d2a55a4fd9/manager/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.273211 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-jdg4h_f1447b71-f9c7-4969-9d8a-d4b8b1f099bb/kube-rbac-proxy/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.313955 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-jdg4h_f1447b71-f9c7-4969-9d8a-d4b8b1f099bb/manager/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.393523 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-wtgcf_c232b1db-4726-4697-b59f-f80c1d9d2d4d/operator/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.511615 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-pknvj_b0e3bc2c-84eb-4407-ae9b-101a70b9d555/kube-rbac-proxy/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.589893 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-m4mrc_b458a451-d4c7-4cb3-b7af-90a12424be9b/kube-rbac-proxy/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.618474 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-pknvj_b0e3bc2c-84eb-4407-ae9b-101a70b9d555/manager/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.735864 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-m4mrc_b458a451-d4c7-4cb3-b7af-90a12424be9b/manager/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.867564 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qj7fr_9b1bcff8-3a58-4c8a-a971-0374ecb16a9b/kube-rbac-proxy/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.890529 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qj7fr_9b1bcff8-3a58-4c8a-a971-0374ecb16a9b/manager/0.log" Dec 06 11:48:27 crc kubenswrapper[4678]: I1206 11:48:27.977725 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-wl7ww_90b306bd-29a2-494d-8399-75d95e9b8597/kube-rbac-proxy/0.log" Dec 06 11:48:28 crc kubenswrapper[4678]: I1206 11:48:28.027716 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-wl7ww_90b306bd-29a2-494d-8399-75d95e9b8597/manager/0.log" Dec 06 11:48:29 crc kubenswrapper[4678]: I1206 11:48:29.505113 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:48:29 crc kubenswrapper[4678]: I1206 11:48:29.505393 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:48:52 crc kubenswrapper[4678]: I1206 11:48:52.125525 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dvssp_9b0941dc-500a-4284-9cc4-66b250feee93/control-plane-machine-set-operator/0.log" Dec 06 11:48:52 crc kubenswrapper[4678]: I1206 11:48:52.231792 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r5rng_0061f615-2fb9-4e17-8ba5-62d97e8070c8/machine-api-operator/0.log" Dec 06 11:48:52 crc kubenswrapper[4678]: I1206 11:48:52.281578 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r5rng_0061f615-2fb9-4e17-8ba5-62d97e8070c8/kube-rbac-proxy/0.log" Dec 06 11:48:59 crc kubenswrapper[4678]: I1206 11:48:59.504979 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:48:59 crc kubenswrapper[4678]: I1206 11:48:59.505600 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:49:07 crc kubenswrapper[4678]: I1206 11:49:07.355617 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-ckqhf_048fff94-2a2e-4819-af68-f3e8b2f2fa3f/cert-manager-controller/0.log" Dec 06 11:49:07 crc kubenswrapper[4678]: I1206 11:49:07.377776 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-8889z_882b9c0d-2380-42b7-b13f-480a4f53a258/cert-manager-cainjector/0.log" Dec 06 11:49:07 crc kubenswrapper[4678]: I1206 11:49:07.914249 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rhwnp_5e3d90b8-1cbb-4329-8d55-911e6dad3243/cert-manager-webhook/0.log" Dec 06 11:49:21 crc kubenswrapper[4678]: I1206 11:49:21.044799 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g72ds_d99e972f-a316-4107-85b4-bf5411999e91/nmstate-console-plugin/0.log" Dec 06 11:49:21 crc kubenswrapper[4678]: I1206 11:49:21.288235 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-n2j4m_82398c4a-8182-4723-bad7-0c289be9d283/nmstate-handler/0.log" Dec 06 11:49:21 crc kubenswrapper[4678]: I1206 11:49:21.306626 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-7nc8k_1e102849-cdd0-4035-a9ba-339d77bd6f8e/kube-rbac-proxy/0.log" Dec 06 11:49:21 crc kubenswrapper[4678]: I1206 11:49:21.401771 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-7nc8k_1e102849-cdd0-4035-a9ba-339d77bd6f8e/nmstate-metrics/0.log" Dec 06 11:49:21 crc kubenswrapper[4678]: I1206 11:49:21.563645 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-5rqmh_fd2552d3-a5fd-4b13-8d01-996a25348826/nmstate-operator/0.log" Dec 06 11:49:21 crc kubenswrapper[4678]: I1206 11:49:21.662206 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-8gwjz_ca2c68b1-3733-4be8-9580-616e9d436a34/nmstate-webhook/0.log" Dec 06 11:49:29 crc kubenswrapper[4678]: I1206 11:49:29.506102 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:49:29 crc kubenswrapper[4678]: I1206 11:49:29.506659 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:49:29 crc kubenswrapper[4678]: I1206 11:49:29.506709 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:49:29 crc kubenswrapper[4678]: I1206 11:49:29.507540 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:49:29 crc kubenswrapper[4678]: I1206 11:49:29.507610 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" gracePeriod=600 Dec 06 11:49:29 crc kubenswrapper[4678]: E1206 11:49:29.628324 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:49:30 crc kubenswrapper[4678]: I1206 11:49:30.286366 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" exitCode=0 Dec 06 11:49:30 crc kubenswrapper[4678]: I1206 11:49:30.286418 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628"} Dec 06 11:49:30 crc kubenswrapper[4678]: I1206 11:49:30.286457 4678 scope.go:117] "RemoveContainer" containerID="841eeeda6a566780d59bc42c1531aafed9b91726537c5ad525315b37460eacdd" Dec 06 11:49:30 crc kubenswrapper[4678]: I1206 11:49:30.287319 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:49:30 crc kubenswrapper[4678]: E1206 11:49:30.287769 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:49:39 crc kubenswrapper[4678]: I1206 11:49:39.946339 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-wstfq_714f1cf8-6f0e-4c0f-8924-4b07a98c578f/kube-rbac-proxy/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.040236 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-wstfq_714f1cf8-6f0e-4c0f-8924-4b07a98c578f/controller/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.160816 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.459703 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.486532 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.487259 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.533307 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.887596 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:49:40 crc kubenswrapper[4678]: I1206 11:49:40.887954 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.153763 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.161624 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.324395 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.337861 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.430837 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.450352 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/controller/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.562535 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/frr-metrics/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.709641 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/kube-rbac-proxy/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.751060 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/kube-rbac-proxy-frr/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.840628 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/reloader/0.log" Dec 06 11:49:41 crc kubenswrapper[4678]: I1206 11:49:41.994886 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-sqk7m_1582c836-cc2c-4d01-b878-6fd5cde0ae70/frr-k8s-webhook-server/0.log" Dec 06 11:49:42 crc kubenswrapper[4678]: I1206 11:49:42.242241 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-dd988988c-fc549_bc881548-41a9-4790-b691-dfb5935d128f/manager/0.log" Dec 06 11:49:42 crc kubenswrapper[4678]: I1206 11:49:42.415557 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d6f755774-dghr8_2562672a-fcdf-420a-a4df-116113c240a0/webhook-server/0.log" Dec 06 11:49:42 crc kubenswrapper[4678]: I1206 11:49:42.782025 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qjlkq_e6542baf-28b2-4844-991e-eb0f57d45231/kube-rbac-proxy/0.log" Dec 06 11:49:42 crc kubenswrapper[4678]: I1206 11:49:42.859386 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/frr/0.log" Dec 06 11:49:43 crc kubenswrapper[4678]: I1206 11:49:43.026143 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qjlkq_e6542baf-28b2-4844-991e-eb0f57d45231/speaker/0.log" Dec 06 11:49:45 crc kubenswrapper[4678]: I1206 11:49:45.479515 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:49:45 crc kubenswrapper[4678]: E1206 11:49:45.480096 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:49:57 crc kubenswrapper[4678]: I1206 11:49:57.983089 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/util/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.148551 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/pull/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.196225 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/util/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.220152 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/pull/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.383453 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/pull/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.392444 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/extract/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.427264 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/util/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.475434 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:49:58 crc kubenswrapper[4678]: E1206 11:49:58.475685 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.581632 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/util/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.810550 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/pull/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.813576 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/pull/0.log" Dec 06 11:49:58 crc kubenswrapper[4678]: I1206 11:49:58.852105 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/util/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.050880 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/util/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.060308 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/pull/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.126370 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/extract/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.297268 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-utilities/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.461092 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-content/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.540787 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-utilities/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.587066 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-content/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.732029 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-utilities/0.log" Dec 06 11:49:59 crc kubenswrapper[4678]: I1206 11:49:59.819994 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-content/0.log" Dec 06 11:50:00 crc kubenswrapper[4678]: I1206 11:50:00.118574 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-utilities/0.log" Dec 06 11:50:00 crc kubenswrapper[4678]: I1206 11:50:00.287391 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/registry-server/0.log" Dec 06 11:50:00 crc kubenswrapper[4678]: I1206 11:50:00.355098 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-content/0.log" Dec 06 11:50:00 crc kubenswrapper[4678]: I1206 11:50:00.374150 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-utilities/0.log" Dec 06 11:50:00 crc kubenswrapper[4678]: I1206 11:50:00.414619 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-content/0.log" Dec 06 11:50:00 crc kubenswrapper[4678]: I1206 11:50:00.660317 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-utilities/0.log" Dec 06 11:50:00 crc kubenswrapper[4678]: I1206 11:50:00.671618 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-content/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.114612 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-s46l9_7dc2165b-f6c0-4386-b931-5e80f762dab5/marketplace-operator/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.198199 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/registry-server/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.233503 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-utilities/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.506036 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-content/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.527834 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-content/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.529674 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-utilities/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.671004 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-utilities/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.682319 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-content/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.818429 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/registry-server/0.log" Dec 06 11:50:01 crc kubenswrapper[4678]: I1206 11:50:01.907796 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-utilities/0.log" Dec 06 11:50:02 crc kubenswrapper[4678]: I1206 11:50:02.155782 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-content/0.log" Dec 06 11:50:02 crc kubenswrapper[4678]: I1206 11:50:02.165634 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-utilities/0.log" Dec 06 11:50:02 crc kubenswrapper[4678]: I1206 11:50:02.167800 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-content/0.log" Dec 06 11:50:02 crc kubenswrapper[4678]: I1206 11:50:02.318038 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-utilities/0.log" Dec 06 11:50:02 crc kubenswrapper[4678]: I1206 11:50:02.358557 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-content/0.log" Dec 06 11:50:02 crc kubenswrapper[4678]: I1206 11:50:02.875359 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/registry-server/0.log" Dec 06 11:50:10 crc kubenswrapper[4678]: I1206 11:50:10.475958 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:50:10 crc kubenswrapper[4678]: E1206 11:50:10.476784 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:50:24 crc kubenswrapper[4678]: I1206 11:50:24.476247 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:50:24 crc kubenswrapper[4678]: E1206 11:50:24.477103 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:50:38 crc kubenswrapper[4678]: I1206 11:50:38.475682 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:50:38 crc kubenswrapper[4678]: E1206 11:50:38.476351 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:50:39 crc kubenswrapper[4678]: E1206 11:50:39.675947 4678 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.198:60642->38.102.83.198:33499: write tcp 38.102.83.198:60642->38.102.83.198:33499: write: broken pipe Dec 06 11:50:51 crc kubenswrapper[4678]: I1206 11:50:51.476605 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:50:51 crc kubenswrapper[4678]: E1206 11:50:51.477799 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.003061 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4brlm"] Dec 06 11:50:57 crc kubenswrapper[4678]: E1206 11:50:57.016867 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06285c50-7a5c-4a46-9b8d-e42b26f1a088" containerName="container-00" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.016901 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="06285c50-7a5c-4a46-9b8d-e42b26f1a088" containerName="container-00" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.017232 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="06285c50-7a5c-4a46-9b8d-e42b26f1a088" containerName="container-00" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.018919 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4brlm"] Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.019022 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.129370 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxm99\" (UniqueName: \"kubernetes.io/projected/6959cdd4-8392-40d1-97d3-901c8c654f23-kube-api-access-rxm99\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.129535 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-utilities\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.129609 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-catalog-content\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.231236 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxm99\" (UniqueName: \"kubernetes.io/projected/6959cdd4-8392-40d1-97d3-901c8c654f23-kube-api-access-rxm99\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.231525 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-utilities\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.231655 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-catalog-content\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.232006 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-utilities\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.232059 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-catalog-content\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.268531 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxm99\" (UniqueName: \"kubernetes.io/projected/6959cdd4-8392-40d1-97d3-901c8c654f23-kube-api-access-rxm99\") pod \"community-operators-4brlm\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.337697 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:50:57 crc kubenswrapper[4678]: I1206 11:50:57.887332 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4brlm"] Dec 06 11:50:58 crc kubenswrapper[4678]: I1206 11:50:58.171176 4678 generic.go:334] "Generic (PLEG): container finished" podID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerID="d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248" exitCode=0 Dec 06 11:50:58 crc kubenswrapper[4678]: I1206 11:50:58.171438 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4brlm" event={"ID":"6959cdd4-8392-40d1-97d3-901c8c654f23","Type":"ContainerDied","Data":"d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248"} Dec 06 11:50:58 crc kubenswrapper[4678]: I1206 11:50:58.171470 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4brlm" event={"ID":"6959cdd4-8392-40d1-97d3-901c8c654f23","Type":"ContainerStarted","Data":"0f869445bed80b2ddd64a30d0d6b07cfa27550ce43511ff116983712bfbc0129"} Dec 06 11:50:58 crc kubenswrapper[4678]: I1206 11:50:58.174902 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:50:59 crc kubenswrapper[4678]: I1206 11:50:59.182365 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4brlm" event={"ID":"6959cdd4-8392-40d1-97d3-901c8c654f23","Type":"ContainerStarted","Data":"a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b"} Dec 06 11:51:00 crc kubenswrapper[4678]: I1206 11:51:00.230283 4678 generic.go:334] "Generic (PLEG): container finished" podID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerID="a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b" exitCode=0 Dec 06 11:51:00 crc kubenswrapper[4678]: I1206 11:51:00.230358 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4brlm" event={"ID":"6959cdd4-8392-40d1-97d3-901c8c654f23","Type":"ContainerDied","Data":"a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b"} Dec 06 11:51:02 crc kubenswrapper[4678]: I1206 11:51:02.261979 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4brlm" event={"ID":"6959cdd4-8392-40d1-97d3-901c8c654f23","Type":"ContainerStarted","Data":"776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef"} Dec 06 11:51:02 crc kubenswrapper[4678]: I1206 11:51:02.292751 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4brlm" podStartSLOduration=3.836371223 podStartE2EDuration="6.292725529s" podCreationTimestamp="2025-12-06 11:50:56 +0000 UTC" firstStartedPulling="2025-12-06 11:50:58.174705828 +0000 UTC m=+4463.018137267" lastFinishedPulling="2025-12-06 11:51:00.631060124 +0000 UTC m=+4465.474491573" observedRunningTime="2025-12-06 11:51:02.285077762 +0000 UTC m=+4467.128509201" watchObservedRunningTime="2025-12-06 11:51:02.292725529 +0000 UTC m=+4467.136156988" Dec 06 11:51:03 crc kubenswrapper[4678]: I1206 11:51:03.480629 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:51:03 crc kubenswrapper[4678]: E1206 11:51:03.492352 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:51:07 crc kubenswrapper[4678]: I1206 11:51:07.338559 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:51:07 crc kubenswrapper[4678]: I1206 11:51:07.340656 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:51:07 crc kubenswrapper[4678]: I1206 11:51:07.395831 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:51:08 crc kubenswrapper[4678]: I1206 11:51:08.384700 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:51:08 crc kubenswrapper[4678]: I1206 11:51:08.433866 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4brlm"] Dec 06 11:51:10 crc kubenswrapper[4678]: I1206 11:51:10.331706 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4brlm" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="registry-server" containerID="cri-o://776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef" gracePeriod=2 Dec 06 11:51:10 crc kubenswrapper[4678]: I1206 11:51:10.829076 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:51:10 crc kubenswrapper[4678]: I1206 11:51:10.954202 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-catalog-content\") pod \"6959cdd4-8392-40d1-97d3-901c8c654f23\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " Dec 06 11:51:10 crc kubenswrapper[4678]: I1206 11:51:10.954668 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxm99\" (UniqueName: \"kubernetes.io/projected/6959cdd4-8392-40d1-97d3-901c8c654f23-kube-api-access-rxm99\") pod \"6959cdd4-8392-40d1-97d3-901c8c654f23\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " Dec 06 11:51:10 crc kubenswrapper[4678]: I1206 11:51:10.955781 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-utilities\") pod \"6959cdd4-8392-40d1-97d3-901c8c654f23\" (UID: \"6959cdd4-8392-40d1-97d3-901c8c654f23\") " Dec 06 11:51:10 crc kubenswrapper[4678]: I1206 11:51:10.957116 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-utilities" (OuterVolumeSpecName: "utilities") pod "6959cdd4-8392-40d1-97d3-901c8c654f23" (UID: "6959cdd4-8392-40d1-97d3-901c8c654f23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:51:10 crc kubenswrapper[4678]: I1206 11:51:10.959790 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6959cdd4-8392-40d1-97d3-901c8c654f23-kube-api-access-rxm99" (OuterVolumeSpecName: "kube-api-access-rxm99") pod "6959cdd4-8392-40d1-97d3-901c8c654f23" (UID: "6959cdd4-8392-40d1-97d3-901c8c654f23"). InnerVolumeSpecName "kube-api-access-rxm99". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.037811 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6959cdd4-8392-40d1-97d3-901c8c654f23" (UID: "6959cdd4-8392-40d1-97d3-901c8c654f23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.059605 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.059639 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6959cdd4-8392-40d1-97d3-901c8c654f23-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.059650 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxm99\" (UniqueName: \"kubernetes.io/projected/6959cdd4-8392-40d1-97d3-901c8c654f23-kube-api-access-rxm99\") on node \"crc\" DevicePath \"\"" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.344441 4678 generic.go:334] "Generic (PLEG): container finished" podID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerID="776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef" exitCode=0 Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.344481 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4brlm" event={"ID":"6959cdd4-8392-40d1-97d3-901c8c654f23","Type":"ContainerDied","Data":"776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef"} Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.344528 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4brlm" event={"ID":"6959cdd4-8392-40d1-97d3-901c8c654f23","Type":"ContainerDied","Data":"0f869445bed80b2ddd64a30d0d6b07cfa27550ce43511ff116983712bfbc0129"} Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.344546 4678 scope.go:117] "RemoveContainer" containerID="776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.344599 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4brlm" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.397722 4678 scope.go:117] "RemoveContainer" containerID="a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.401786 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4brlm"] Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.413309 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4brlm"] Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.486632 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" path="/var/lib/kubelet/pods/6959cdd4-8392-40d1-97d3-901c8c654f23/volumes" Dec 06 11:51:11 crc kubenswrapper[4678]: I1206 11:51:11.903526 4678 scope.go:117] "RemoveContainer" containerID="d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248" Dec 06 11:51:12 crc kubenswrapper[4678]: I1206 11:51:12.139918 4678 scope.go:117] "RemoveContainer" containerID="776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef" Dec 06 11:51:12 crc kubenswrapper[4678]: E1206 11:51:12.140285 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef\": container with ID starting with 776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef not found: ID does not exist" containerID="776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef" Dec 06 11:51:12 crc kubenswrapper[4678]: I1206 11:51:12.140344 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef"} err="failed to get container status \"776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef\": rpc error: code = NotFound desc = could not find container \"776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef\": container with ID starting with 776f826de983f9281df12bf0d68f48bda69eac89697cfb7f5e10d70b8db47eef not found: ID does not exist" Dec 06 11:51:12 crc kubenswrapper[4678]: I1206 11:51:12.140370 4678 scope.go:117] "RemoveContainer" containerID="a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b" Dec 06 11:51:12 crc kubenswrapper[4678]: E1206 11:51:12.140673 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b\": container with ID starting with a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b not found: ID does not exist" containerID="a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b" Dec 06 11:51:12 crc kubenswrapper[4678]: I1206 11:51:12.140703 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b"} err="failed to get container status \"a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b\": rpc error: code = NotFound desc = could not find container \"a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b\": container with ID starting with a44e4d38300f361e3681885881c09c5f28ec5eeee87d7c1b941fdb6f34a2383b not found: ID does not exist" Dec 06 11:51:12 crc kubenswrapper[4678]: I1206 11:51:12.140720 4678 scope.go:117] "RemoveContainer" containerID="d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248" Dec 06 11:51:12 crc kubenswrapper[4678]: E1206 11:51:12.140973 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248\": container with ID starting with d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248 not found: ID does not exist" containerID="d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248" Dec 06 11:51:12 crc kubenswrapper[4678]: I1206 11:51:12.141000 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248"} err="failed to get container status \"d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248\": rpc error: code = NotFound desc = could not find container \"d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248\": container with ID starting with d930262710123ad82d15ccb193d6fd6d23ec501ca15b3fb2d72c689a46e8f248 not found: ID does not exist" Dec 06 11:51:18 crc kubenswrapper[4678]: I1206 11:51:18.476694 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:51:18 crc kubenswrapper[4678]: E1206 11:51:18.478621 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:51:32 crc kubenswrapper[4678]: I1206 11:51:32.475926 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:51:32 crc kubenswrapper[4678]: E1206 11:51:32.476768 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:51:45 crc kubenswrapper[4678]: I1206 11:51:45.485588 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:51:45 crc kubenswrapper[4678]: E1206 11:51:45.486579 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:51:58 crc kubenswrapper[4678]: I1206 11:51:58.476824 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:51:58 crc kubenswrapper[4678]: E1206 11:51:58.477983 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.149644 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8bxlk"] Dec 06 11:52:06 crc kubenswrapper[4678]: E1206 11:52:06.151710 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="registry-server" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.151815 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="registry-server" Dec 06 11:52:06 crc kubenswrapper[4678]: E1206 11:52:06.151888 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="extract-content" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.151957 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="extract-content" Dec 06 11:52:06 crc kubenswrapper[4678]: E1206 11:52:06.152061 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="extract-utilities" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.152140 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="extract-utilities" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.152450 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6959cdd4-8392-40d1-97d3-901c8c654f23" containerName="registry-server" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.154292 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.163948 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bxlk"] Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.194243 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-utilities\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.194347 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-catalog-content\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.194401 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmjwz\" (UniqueName: \"kubernetes.io/projected/e114c992-2e7b-498c-a927-a6597fbd75d8-kube-api-access-lmjwz\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.296467 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmjwz\" (UniqueName: \"kubernetes.io/projected/e114c992-2e7b-498c-a927-a6597fbd75d8-kube-api-access-lmjwz\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.296605 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-utilities\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.296729 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-catalog-content\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.297347 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-catalog-content\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.297547 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-utilities\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.319538 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmjwz\" (UniqueName: \"kubernetes.io/projected/e114c992-2e7b-498c-a927-a6597fbd75d8-kube-api-access-lmjwz\") pod \"redhat-marketplace-8bxlk\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.362921 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rgk7n"] Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.364912 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.371693 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rgk7n"] Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.398121 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n72gb\" (UniqueName: \"kubernetes.io/projected/bc02324f-d15e-42c6-9111-e1be7663010b-kube-api-access-n72gb\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.398181 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-catalog-content\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.398335 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-utilities\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.475567 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.499552 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-catalog-content\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.499905 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-utilities\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.500006 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-catalog-content\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.500986 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-utilities\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.501310 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n72gb\" (UniqueName: \"kubernetes.io/projected/bc02324f-d15e-42c6-9111-e1be7663010b-kube-api-access-n72gb\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.536351 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n72gb\" (UniqueName: \"kubernetes.io/projected/bc02324f-d15e-42c6-9111-e1be7663010b-kube-api-access-n72gb\") pod \"redhat-operators-rgk7n\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:06 crc kubenswrapper[4678]: I1206 11:52:06.689311 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.080277 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bxlk"] Dec 06 11:52:07 crc kubenswrapper[4678]: W1206 11:52:07.084524 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode114c992_2e7b_498c_a927_a6597fbd75d8.slice/crio-4f69a97b793cd934b0de7cb91895b0f3b91c50b9fefc2e57afc3fb94a72406e0 WatchSource:0}: Error finding container 4f69a97b793cd934b0de7cb91895b0f3b91c50b9fefc2e57afc3fb94a72406e0: Status 404 returned error can't find the container with id 4f69a97b793cd934b0de7cb91895b0f3b91c50b9fefc2e57afc3fb94a72406e0 Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.129339 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rgk7n"] Dec 06 11:52:07 crc kubenswrapper[4678]: W1206 11:52:07.147445 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc02324f_d15e_42c6_9111_e1be7663010b.slice/crio-a61d0911ca7e19cd5f7b1fc262514bc0d7e854c6c93f206fc2b5491d95c98ed8 WatchSource:0}: Error finding container a61d0911ca7e19cd5f7b1fc262514bc0d7e854c6c93f206fc2b5491d95c98ed8: Status 404 returned error can't find the container with id a61d0911ca7e19cd5f7b1fc262514bc0d7e854c6c93f206fc2b5491d95c98ed8 Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.981799 4678 generic.go:334] "Generic (PLEG): container finished" podID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerID="9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc" exitCode=0 Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.981921 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bxlk" event={"ID":"e114c992-2e7b-498c-a927-a6597fbd75d8","Type":"ContainerDied","Data":"9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc"} Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.982386 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bxlk" event={"ID":"e114c992-2e7b-498c-a927-a6597fbd75d8","Type":"ContainerStarted","Data":"4f69a97b793cd934b0de7cb91895b0f3b91c50b9fefc2e57afc3fb94a72406e0"} Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.985215 4678 generic.go:334] "Generic (PLEG): container finished" podID="bc02324f-d15e-42c6-9111-e1be7663010b" containerID="a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e" exitCode=0 Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.985245 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgk7n" event={"ID":"bc02324f-d15e-42c6-9111-e1be7663010b","Type":"ContainerDied","Data":"a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e"} Dec 06 11:52:07 crc kubenswrapper[4678]: I1206 11:52:07.985260 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgk7n" event={"ID":"bc02324f-d15e-42c6-9111-e1be7663010b","Type":"ContainerStarted","Data":"a61d0911ca7e19cd5f7b1fc262514bc0d7e854c6c93f206fc2b5491d95c98ed8"} Dec 06 11:52:09 crc kubenswrapper[4678]: I1206 11:52:08.999733 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bxlk" event={"ID":"e114c992-2e7b-498c-a927-a6597fbd75d8","Type":"ContainerStarted","Data":"6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375"} Dec 06 11:52:09 crc kubenswrapper[4678]: I1206 11:52:09.002264 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgk7n" event={"ID":"bc02324f-d15e-42c6-9111-e1be7663010b","Type":"ContainerStarted","Data":"af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9"} Dec 06 11:52:10 crc kubenswrapper[4678]: I1206 11:52:10.011892 4678 generic.go:334] "Generic (PLEG): container finished" podID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerID="6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375" exitCode=0 Dec 06 11:52:10 crc kubenswrapper[4678]: I1206 11:52:10.011953 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bxlk" event={"ID":"e114c992-2e7b-498c-a927-a6597fbd75d8","Type":"ContainerDied","Data":"6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375"} Dec 06 11:52:12 crc kubenswrapper[4678]: I1206 11:52:12.483597 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:52:12 crc kubenswrapper[4678]: E1206 11:52:12.485793 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:52:13 crc kubenswrapper[4678]: I1206 11:52:13.041876 4678 generic.go:334] "Generic (PLEG): container finished" podID="bc02324f-d15e-42c6-9111-e1be7663010b" containerID="af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9" exitCode=0 Dec 06 11:52:13 crc kubenswrapper[4678]: I1206 11:52:13.041910 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgk7n" event={"ID":"bc02324f-d15e-42c6-9111-e1be7663010b","Type":"ContainerDied","Data":"af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9"} Dec 06 11:52:13 crc kubenswrapper[4678]: I1206 11:52:13.045327 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bxlk" event={"ID":"e114c992-2e7b-498c-a927-a6597fbd75d8","Type":"ContainerStarted","Data":"7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159"} Dec 06 11:52:13 crc kubenswrapper[4678]: I1206 11:52:13.106293 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8bxlk" podStartSLOduration=2.833518369 podStartE2EDuration="7.10627266s" podCreationTimestamp="2025-12-06 11:52:06 +0000 UTC" firstStartedPulling="2025-12-06 11:52:07.984987018 +0000 UTC m=+4532.828418447" lastFinishedPulling="2025-12-06 11:52:12.257741289 +0000 UTC m=+4537.101172738" observedRunningTime="2025-12-06 11:52:13.105984842 +0000 UTC m=+4537.949416291" watchObservedRunningTime="2025-12-06 11:52:13.10627266 +0000 UTC m=+4537.949704089" Dec 06 11:52:15 crc kubenswrapper[4678]: I1206 11:52:15.062442 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgk7n" event={"ID":"bc02324f-d15e-42c6-9111-e1be7663010b","Type":"ContainerStarted","Data":"88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2"} Dec 06 11:52:15 crc kubenswrapper[4678]: I1206 11:52:15.086173 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rgk7n" podStartSLOduration=3.6348039549999998 podStartE2EDuration="9.086153146s" podCreationTimestamp="2025-12-06 11:52:06 +0000 UTC" firstStartedPulling="2025-12-06 11:52:07.986515175 +0000 UTC m=+4532.829946614" lastFinishedPulling="2025-12-06 11:52:13.437864366 +0000 UTC m=+4538.281295805" observedRunningTime="2025-12-06 11:52:15.082894125 +0000 UTC m=+4539.926325564" watchObservedRunningTime="2025-12-06 11:52:15.086153146 +0000 UTC m=+4539.929584585" Dec 06 11:52:16 crc kubenswrapper[4678]: I1206 11:52:16.476379 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:16 crc kubenswrapper[4678]: I1206 11:52:16.476816 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:16 crc kubenswrapper[4678]: I1206 11:52:16.538261 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:16 crc kubenswrapper[4678]: I1206 11:52:16.691066 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:16 crc kubenswrapper[4678]: I1206 11:52:16.691304 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:17 crc kubenswrapper[4678]: I1206 11:52:17.146119 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:17 crc kubenswrapper[4678]: I1206 11:52:17.749190 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rgk7n" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="registry-server" probeResult="failure" output=< Dec 06 11:52:17 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 11:52:17 crc kubenswrapper[4678]: > Dec 06 11:52:17 crc kubenswrapper[4678]: I1206 11:52:17.938971 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bxlk"] Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.095996 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8bxlk" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="registry-server" containerID="cri-o://7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159" gracePeriod=2 Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.565241 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.701548 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-catalog-content\") pod \"e114c992-2e7b-498c-a927-a6597fbd75d8\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.701728 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmjwz\" (UniqueName: \"kubernetes.io/projected/e114c992-2e7b-498c-a927-a6597fbd75d8-kube-api-access-lmjwz\") pod \"e114c992-2e7b-498c-a927-a6597fbd75d8\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.701813 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-utilities\") pod \"e114c992-2e7b-498c-a927-a6597fbd75d8\" (UID: \"e114c992-2e7b-498c-a927-a6597fbd75d8\") " Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.702421 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-utilities" (OuterVolumeSpecName: "utilities") pod "e114c992-2e7b-498c-a927-a6597fbd75d8" (UID: "e114c992-2e7b-498c-a927-a6597fbd75d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.707478 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e114c992-2e7b-498c-a927-a6597fbd75d8-kube-api-access-lmjwz" (OuterVolumeSpecName: "kube-api-access-lmjwz") pod "e114c992-2e7b-498c-a927-a6597fbd75d8" (UID: "e114c992-2e7b-498c-a927-a6597fbd75d8"). InnerVolumeSpecName "kube-api-access-lmjwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.726814 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e114c992-2e7b-498c-a927-a6597fbd75d8" (UID: "e114c992-2e7b-498c-a927-a6597fbd75d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.804784 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.804829 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmjwz\" (UniqueName: \"kubernetes.io/projected/e114c992-2e7b-498c-a927-a6597fbd75d8-kube-api-access-lmjwz\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:19 crc kubenswrapper[4678]: I1206 11:52:19.804853 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e114c992-2e7b-498c-a927-a6597fbd75d8-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.106365 4678 generic.go:334] "Generic (PLEG): container finished" podID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerID="7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159" exitCode=0 Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.106430 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bxlk" event={"ID":"e114c992-2e7b-498c-a927-a6597fbd75d8","Type":"ContainerDied","Data":"7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159"} Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.106455 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8bxlk" event={"ID":"e114c992-2e7b-498c-a927-a6597fbd75d8","Type":"ContainerDied","Data":"4f69a97b793cd934b0de7cb91895b0f3b91c50b9fefc2e57afc3fb94a72406e0"} Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.106480 4678 scope.go:117] "RemoveContainer" containerID="7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.106606 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8bxlk" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.164374 4678 scope.go:117] "RemoveContainer" containerID="6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.165454 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bxlk"] Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.175155 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8bxlk"] Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.191661 4678 scope.go:117] "RemoveContainer" containerID="9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.230786 4678 scope.go:117] "RemoveContainer" containerID="7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159" Dec 06 11:52:20 crc kubenswrapper[4678]: E1206 11:52:20.231608 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159\": container with ID starting with 7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159 not found: ID does not exist" containerID="7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.231664 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159"} err="failed to get container status \"7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159\": rpc error: code = NotFound desc = could not find container \"7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159\": container with ID starting with 7b6114ae3556a988e51aaccd589393879ce4c55af00740702957f1d9e7cd6159 not found: ID does not exist" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.231690 4678 scope.go:117] "RemoveContainer" containerID="6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375" Dec 06 11:52:20 crc kubenswrapper[4678]: E1206 11:52:20.232308 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375\": container with ID starting with 6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375 not found: ID does not exist" containerID="6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.232334 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375"} err="failed to get container status \"6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375\": rpc error: code = NotFound desc = could not find container \"6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375\": container with ID starting with 6e78ce565cd11a8f00ba1c0c9efbc56edc30550b690188f4356a3276d039c375 not found: ID does not exist" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.232347 4678 scope.go:117] "RemoveContainer" containerID="9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc" Dec 06 11:52:20 crc kubenswrapper[4678]: E1206 11:52:20.233111 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc\": container with ID starting with 9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc not found: ID does not exist" containerID="9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc" Dec 06 11:52:20 crc kubenswrapper[4678]: I1206 11:52:20.233144 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc"} err="failed to get container status \"9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc\": rpc error: code = NotFound desc = could not find container \"9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc\": container with ID starting with 9abd51e72509e878dff8e64b6e21f3ea09af3adc7b494d358f4442593f66e1bc not found: ID does not exist" Dec 06 11:52:21 crc kubenswrapper[4678]: I1206 11:52:21.490784 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" path="/var/lib/kubelet/pods/e114c992-2e7b-498c-a927-a6597fbd75d8/volumes" Dec 06 11:52:23 crc kubenswrapper[4678]: I1206 11:52:23.139788 4678 generic.go:334] "Generic (PLEG): container finished" podID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerID="c796f2dad127e1fea701eecab88d92e927db3b4d9d4b36c2502ab66c32fa4602" exitCode=0 Dec 06 11:52:23 crc kubenswrapper[4678]: I1206 11:52:23.140592 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" event={"ID":"af614e30-ef67-4ebc-bf1c-7203ba99b625","Type":"ContainerDied","Data":"c796f2dad127e1fea701eecab88d92e927db3b4d9d4b36c2502ab66c32fa4602"} Dec 06 11:52:23 crc kubenswrapper[4678]: I1206 11:52:23.141191 4678 scope.go:117] "RemoveContainer" containerID="c796f2dad127e1fea701eecab88d92e927db3b4d9d4b36c2502ab66c32fa4602" Dec 06 11:52:23 crc kubenswrapper[4678]: I1206 11:52:23.477823 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:52:23 crc kubenswrapper[4678]: E1206 11:52:23.478082 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:52:24 crc kubenswrapper[4678]: I1206 11:52:24.133909 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ngf8_must-gather-xqtv2_af614e30-ef67-4ebc-bf1c-7203ba99b625/gather/0.log" Dec 06 11:52:26 crc kubenswrapper[4678]: I1206 11:52:26.813045 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:26 crc kubenswrapper[4678]: I1206 11:52:26.876220 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:27 crc kubenswrapper[4678]: I1206 11:52:27.059315 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rgk7n"] Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.185663 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rgk7n" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="registry-server" containerID="cri-o://88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2" gracePeriod=2 Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.706797 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.825888 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n72gb\" (UniqueName: \"kubernetes.io/projected/bc02324f-d15e-42c6-9111-e1be7663010b-kube-api-access-n72gb\") pod \"bc02324f-d15e-42c6-9111-e1be7663010b\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.825982 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-utilities\") pod \"bc02324f-d15e-42c6-9111-e1be7663010b\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.826094 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-catalog-content\") pod \"bc02324f-d15e-42c6-9111-e1be7663010b\" (UID: \"bc02324f-d15e-42c6-9111-e1be7663010b\") " Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.826656 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-utilities" (OuterVolumeSpecName: "utilities") pod "bc02324f-d15e-42c6-9111-e1be7663010b" (UID: "bc02324f-d15e-42c6-9111-e1be7663010b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.832075 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc02324f-d15e-42c6-9111-e1be7663010b-kube-api-access-n72gb" (OuterVolumeSpecName: "kube-api-access-n72gb") pod "bc02324f-d15e-42c6-9111-e1be7663010b" (UID: "bc02324f-d15e-42c6-9111-e1be7663010b"). InnerVolumeSpecName "kube-api-access-n72gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.928104 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n72gb\" (UniqueName: \"kubernetes.io/projected/bc02324f-d15e-42c6-9111-e1be7663010b-kube-api-access-n72gb\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.928135 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:28 crc kubenswrapper[4678]: I1206 11:52:28.945693 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc02324f-d15e-42c6-9111-e1be7663010b" (UID: "bc02324f-d15e-42c6-9111-e1be7663010b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.030294 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc02324f-d15e-42c6-9111-e1be7663010b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.197513 4678 generic.go:334] "Generic (PLEG): container finished" podID="bc02324f-d15e-42c6-9111-e1be7663010b" containerID="88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2" exitCode=0 Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.197563 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgk7n" event={"ID":"bc02324f-d15e-42c6-9111-e1be7663010b","Type":"ContainerDied","Data":"88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2"} Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.197602 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rgk7n" event={"ID":"bc02324f-d15e-42c6-9111-e1be7663010b","Type":"ContainerDied","Data":"a61d0911ca7e19cd5f7b1fc262514bc0d7e854c6c93f206fc2b5491d95c98ed8"} Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.197635 4678 scope.go:117] "RemoveContainer" containerID="88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.197681 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rgk7n" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.231310 4678 scope.go:117] "RemoveContainer" containerID="af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.273708 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rgk7n"] Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.290093 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rgk7n"] Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.291958 4678 scope.go:117] "RemoveContainer" containerID="a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.314453 4678 scope.go:117] "RemoveContainer" containerID="88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2" Dec 06 11:52:29 crc kubenswrapper[4678]: E1206 11:52:29.315766 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2\": container with ID starting with 88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2 not found: ID does not exist" containerID="88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.315822 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2"} err="failed to get container status \"88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2\": rpc error: code = NotFound desc = could not find container \"88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2\": container with ID starting with 88e7ce480d0e13019f54432b4f63848c4dbfa26b87fadc4733ebe914d4c273e2 not found: ID does not exist" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.315853 4678 scope.go:117] "RemoveContainer" containerID="af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9" Dec 06 11:52:29 crc kubenswrapper[4678]: E1206 11:52:29.317452 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9\": container with ID starting with af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9 not found: ID does not exist" containerID="af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.317585 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9"} err="failed to get container status \"af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9\": rpc error: code = NotFound desc = could not find container \"af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9\": container with ID starting with af92a71d72e4a489c93338797c8020316fc3e74aa6b5363dbab462ef95f7d9d9 not found: ID does not exist" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.317611 4678 scope.go:117] "RemoveContainer" containerID="a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e" Dec 06 11:52:29 crc kubenswrapper[4678]: E1206 11:52:29.318103 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e\": container with ID starting with a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e not found: ID does not exist" containerID="a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.318777 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e"} err="failed to get container status \"a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e\": rpc error: code = NotFound desc = could not find container \"a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e\": container with ID starting with a53a86aadda02c0ed96a7bfba639fe953dc630d7d897881c30200c429278da9e not found: ID does not exist" Dec 06 11:52:29 crc kubenswrapper[4678]: I1206 11:52:29.487226 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" path="/var/lib/kubelet/pods/bc02324f-d15e-42c6-9111-e1be7663010b/volumes" Dec 06 11:52:32 crc kubenswrapper[4678]: I1206 11:52:32.933618 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-8ngf8/must-gather-xqtv2"] Dec 06 11:52:32 crc kubenswrapper[4678]: I1206 11:52:32.934384 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerName="copy" containerID="cri-o://b453a8c10e76d33366016dca59cb9ab158f42dc8394a0aa13ece77400883b35c" gracePeriod=2 Dec 06 11:52:32 crc kubenswrapper[4678]: I1206 11:52:32.943434 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-8ngf8/must-gather-xqtv2"] Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.247157 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ngf8_must-gather-xqtv2_af614e30-ef67-4ebc-bf1c-7203ba99b625/copy/0.log" Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.248274 4678 generic.go:334] "Generic (PLEG): container finished" podID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerID="b453a8c10e76d33366016dca59cb9ab158f42dc8394a0aa13ece77400883b35c" exitCode=143 Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.385303 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ngf8_must-gather-xqtv2_af614e30-ef67-4ebc-bf1c-7203ba99b625/copy/0.log" Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.385672 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.524774 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af614e30-ef67-4ebc-bf1c-7203ba99b625-must-gather-output\") pod \"af614e30-ef67-4ebc-bf1c-7203ba99b625\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.524908 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dm686\" (UniqueName: \"kubernetes.io/projected/af614e30-ef67-4ebc-bf1c-7203ba99b625-kube-api-access-dm686\") pod \"af614e30-ef67-4ebc-bf1c-7203ba99b625\" (UID: \"af614e30-ef67-4ebc-bf1c-7203ba99b625\") " Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.531901 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af614e30-ef67-4ebc-bf1c-7203ba99b625-kube-api-access-dm686" (OuterVolumeSpecName: "kube-api-access-dm686") pod "af614e30-ef67-4ebc-bf1c-7203ba99b625" (UID: "af614e30-ef67-4ebc-bf1c-7203ba99b625"). InnerVolumeSpecName "kube-api-access-dm686". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.626838 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dm686\" (UniqueName: \"kubernetes.io/projected/af614e30-ef67-4ebc-bf1c-7203ba99b625-kube-api-access-dm686\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.691042 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af614e30-ef67-4ebc-bf1c-7203ba99b625-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "af614e30-ef67-4ebc-bf1c-7203ba99b625" (UID: "af614e30-ef67-4ebc-bf1c-7203ba99b625"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:52:33 crc kubenswrapper[4678]: I1206 11:52:33.728244 4678 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/af614e30-ef67-4ebc-bf1c-7203ba99b625-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 11:52:34 crc kubenswrapper[4678]: I1206 11:52:34.259137 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-8ngf8_must-gather-xqtv2_af614e30-ef67-4ebc-bf1c-7203ba99b625/copy/0.log" Dec 06 11:52:34 crc kubenswrapper[4678]: I1206 11:52:34.260383 4678 scope.go:117] "RemoveContainer" containerID="b453a8c10e76d33366016dca59cb9ab158f42dc8394a0aa13ece77400883b35c" Dec 06 11:52:34 crc kubenswrapper[4678]: I1206 11:52:34.260619 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-8ngf8/must-gather-xqtv2" Dec 06 11:52:34 crc kubenswrapper[4678]: I1206 11:52:34.294135 4678 scope.go:117] "RemoveContainer" containerID="c796f2dad127e1fea701eecab88d92e927db3b4d9d4b36c2502ab66c32fa4602" Dec 06 11:52:35 crc kubenswrapper[4678]: I1206 11:52:35.493522 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" path="/var/lib/kubelet/pods/af614e30-ef67-4ebc-bf1c-7203ba99b625/volumes" Dec 06 11:52:38 crc kubenswrapper[4678]: I1206 11:52:38.477548 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:52:38 crc kubenswrapper[4678]: E1206 11:52:38.479084 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:52:48 crc kubenswrapper[4678]: I1206 11:52:48.784058 4678 scope.go:117] "RemoveContainer" containerID="aac116ff92464dbecb1b447113862aca923316d9dae3de79eef655fab045da5c" Dec 06 11:52:48 crc kubenswrapper[4678]: I1206 11:52:48.806016 4678 scope.go:117] "RemoveContainer" containerID="2ae7b1674831dac6a0dd5675b415f1e4e8597d340747778b7ca2e834cae71029" Dec 06 11:52:48 crc kubenswrapper[4678]: I1206 11:52:48.833143 4678 scope.go:117] "RemoveContainer" containerID="4f3e0435b5d82b1deb30ae035af4c71e6b50a0c866b953757f829775a3c2772f" Dec 06 11:52:48 crc kubenswrapper[4678]: I1206 11:52:48.879324 4678 scope.go:117] "RemoveContainer" containerID="27e91936bf33627ce404139dbfa6691e445820fe5cdf7dad9a2862e441c71d00" Dec 06 11:52:51 crc kubenswrapper[4678]: I1206 11:52:51.482933 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:52:51 crc kubenswrapper[4678]: E1206 11:52:51.484732 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:53:06 crc kubenswrapper[4678]: I1206 11:53:06.476546 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:53:06 crc kubenswrapper[4678]: E1206 11:53:06.479004 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:53:19 crc kubenswrapper[4678]: I1206 11:53:19.479612 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:53:19 crc kubenswrapper[4678]: E1206 11:53:19.480479 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:53:33 crc kubenswrapper[4678]: I1206 11:53:33.475708 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:53:33 crc kubenswrapper[4678]: E1206 11:53:33.476660 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:53:46 crc kubenswrapper[4678]: I1206 11:53:46.475693 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:53:46 crc kubenswrapper[4678]: E1206 11:53:46.476730 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:53:49 crc kubenswrapper[4678]: I1206 11:53:49.069181 4678 scope.go:117] "RemoveContainer" containerID="61f56a469e44fd7f45cb7628f32837de6922e4d664bb30f5a8bc2f8cbd4d904d" Dec 06 11:54:01 crc kubenswrapper[4678]: I1206 11:54:01.478306 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:54:01 crc kubenswrapper[4678]: E1206 11:54:01.479061 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:54:15 crc kubenswrapper[4678]: I1206 11:54:15.483998 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:54:15 crc kubenswrapper[4678]: E1206 11:54:15.485773 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:54:28 crc kubenswrapper[4678]: I1206 11:54:28.475960 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:54:28 crc kubenswrapper[4678]: E1206 11:54:28.476913 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 11:54:40 crc kubenswrapper[4678]: I1206 11:54:40.476962 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:54:40 crc kubenswrapper[4678]: I1206 11:54:40.944761 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"a8c3c8b241bd53b83426412a64cbd2cddf1cbe5ae5bf0d01ab9debe1412cf227"} Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.174278 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-shwqz/must-gather-2d8pq"] Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175016 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="extract-utilities" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175027 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="extract-utilities" Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175046 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerName="gather" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175053 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerName="gather" Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175073 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerName="copy" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175079 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerName="copy" Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175090 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="extract-content" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175096 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="extract-content" Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175106 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="extract-content" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175112 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="extract-content" Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175125 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="extract-utilities" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175132 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="extract-utilities" Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175157 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="registry-server" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175163 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="registry-server" Dec 06 11:55:42 crc kubenswrapper[4678]: E1206 11:55:42.175171 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="registry-server" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175177 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="registry-server" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175368 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerName="copy" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175379 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="af614e30-ef67-4ebc-bf1c-7203ba99b625" containerName="gather" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175387 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc02324f-d15e-42c6-9111-e1be7663010b" containerName="registry-server" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.175403 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="e114c992-2e7b-498c-a927-a6597fbd75d8" containerName="registry-server" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.176367 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.185617 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-shwqz"/"default-dockercfg-hkmc6" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.186396 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-shwqz"/"openshift-service-ca.crt" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.186475 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-shwqz"/"kube-root-ca.crt" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.217500 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-shwqz/must-gather-2d8pq"] Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.292022 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aebef180-e4b3-4f57-843c-24d89b64cb14-must-gather-output\") pod \"must-gather-2d8pq\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.292361 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8h6f\" (UniqueName: \"kubernetes.io/projected/aebef180-e4b3-4f57-843c-24d89b64cb14-kube-api-access-k8h6f\") pod \"must-gather-2d8pq\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.394721 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8h6f\" (UniqueName: \"kubernetes.io/projected/aebef180-e4b3-4f57-843c-24d89b64cb14-kube-api-access-k8h6f\") pod \"must-gather-2d8pq\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.395150 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aebef180-e4b3-4f57-843c-24d89b64cb14-must-gather-output\") pod \"must-gather-2d8pq\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.395698 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aebef180-e4b3-4f57-843c-24d89b64cb14-must-gather-output\") pod \"must-gather-2d8pq\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.424730 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8h6f\" (UniqueName: \"kubernetes.io/projected/aebef180-e4b3-4f57-843c-24d89b64cb14-kube-api-access-k8h6f\") pod \"must-gather-2d8pq\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:42 crc kubenswrapper[4678]: I1206 11:55:42.494731 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 11:55:43 crc kubenswrapper[4678]: I1206 11:55:43.044184 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-shwqz/must-gather-2d8pq"] Dec 06 11:55:43 crc kubenswrapper[4678]: I1206 11:55:43.582766 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/must-gather-2d8pq" event={"ID":"aebef180-e4b3-4f57-843c-24d89b64cb14","Type":"ContainerStarted","Data":"499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df"} Dec 06 11:55:43 crc kubenswrapper[4678]: I1206 11:55:43.584088 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/must-gather-2d8pq" event={"ID":"aebef180-e4b3-4f57-843c-24d89b64cb14","Type":"ContainerStarted","Data":"22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86"} Dec 06 11:55:43 crc kubenswrapper[4678]: I1206 11:55:43.584115 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/must-gather-2d8pq" event={"ID":"aebef180-e4b3-4f57-843c-24d89b64cb14","Type":"ContainerStarted","Data":"fd4e98366a2b53f2ea5bda78ef87c815422fbb79d60713b1600844ddceddb465"} Dec 06 11:55:43 crc kubenswrapper[4678]: I1206 11:55:43.600113 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-shwqz/must-gather-2d8pq" podStartSLOduration=1.6000996029999999 podStartE2EDuration="1.600099603s" podCreationTimestamp="2025-12-06 11:55:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:55:43.595027303 +0000 UTC m=+4748.438458742" watchObservedRunningTime="2025-12-06 11:55:43.600099603 +0000 UTC m=+4748.443531042" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.084182 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-shwqz/crc-debug-vtdh2"] Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.086785 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.198120 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-host\") pod \"crc-debug-vtdh2\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.198623 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwwbh\" (UniqueName: \"kubernetes.io/projected/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-kube-api-access-dwwbh\") pod \"crc-debug-vtdh2\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.300350 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-host\") pod \"crc-debug-vtdh2\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.300482 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwwbh\" (UniqueName: \"kubernetes.io/projected/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-kube-api-access-dwwbh\") pod \"crc-debug-vtdh2\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.300882 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-host\") pod \"crc-debug-vtdh2\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.327645 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwwbh\" (UniqueName: \"kubernetes.io/projected/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-kube-api-access-dwwbh\") pod \"crc-debug-vtdh2\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.406711 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:55:47 crc kubenswrapper[4678]: W1206 11:55:47.448882 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e6911ff_f2bb_42d8_9b1b_fd36b2556f0b.slice/crio-143c857050c8ce79203e11b1b1d9928a4c1ca90c0a786e10be3f9d67aa8f85f9 WatchSource:0}: Error finding container 143c857050c8ce79203e11b1b1d9928a4c1ca90c0a786e10be3f9d67aa8f85f9: Status 404 returned error can't find the container with id 143c857050c8ce79203e11b1b1d9928a4c1ca90c0a786e10be3f9d67aa8f85f9 Dec 06 11:55:47 crc kubenswrapper[4678]: I1206 11:55:47.624574 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" event={"ID":"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b","Type":"ContainerStarted","Data":"143c857050c8ce79203e11b1b1d9928a4c1ca90c0a786e10be3f9d67aa8f85f9"} Dec 06 11:55:48 crc kubenswrapper[4678]: I1206 11:55:48.635881 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" event={"ID":"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b","Type":"ContainerStarted","Data":"fab1d58958ee3dcb69ef414b57938f17ab2c9213544cb0e498087e94b966ab8e"} Dec 06 11:55:48 crc kubenswrapper[4678]: I1206 11:55:48.664857 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" podStartSLOduration=1.664834513 podStartE2EDuration="1.664834513s" podCreationTimestamp="2025-12-06 11:55:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:55:48.655516421 +0000 UTC m=+4753.498947860" watchObservedRunningTime="2025-12-06 11:55:48.664834513 +0000 UTC m=+4753.508265942" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.039056 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xfcxd"] Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.041477 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.054066 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfcxd"] Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.190163 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-catalog-content\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.190540 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt5q6\" (UniqueName: \"kubernetes.io/projected/ded25249-eeb5-41f5-b267-32d55af7d2be-kube-api-access-zt5q6\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.190639 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-utilities\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.292273 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-catalog-content\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.292322 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt5q6\" (UniqueName: \"kubernetes.io/projected/ded25249-eeb5-41f5-b267-32d55af7d2be-kube-api-access-zt5q6\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.292378 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-utilities\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.292835 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-catalog-content\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.292898 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-utilities\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.310319 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt5q6\" (UniqueName: \"kubernetes.io/projected/ded25249-eeb5-41f5-b267-32d55af7d2be-kube-api-access-zt5q6\") pod \"certified-operators-xfcxd\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:01 crc kubenswrapper[4678]: I1206 11:56:01.377221 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:02 crc kubenswrapper[4678]: I1206 11:56:02.039013 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xfcxd"] Dec 06 11:56:02 crc kubenswrapper[4678]: I1206 11:56:02.763793 4678 generic.go:334] "Generic (PLEG): container finished" podID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerID="814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08" exitCode=0 Dec 06 11:56:02 crc kubenswrapper[4678]: I1206 11:56:02.763897 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcxd" event={"ID":"ded25249-eeb5-41f5-b267-32d55af7d2be","Type":"ContainerDied","Data":"814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08"} Dec 06 11:56:02 crc kubenswrapper[4678]: I1206 11:56:02.765123 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcxd" event={"ID":"ded25249-eeb5-41f5-b267-32d55af7d2be","Type":"ContainerStarted","Data":"2054400e6cb4045b9eaec6b0e3016819acaf8c6ac9259a420b8bf56ae8bfd5a8"} Dec 06 11:56:02 crc kubenswrapper[4678]: I1206 11:56:02.765849 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 11:56:03 crc kubenswrapper[4678]: I1206 11:56:03.776011 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcxd" event={"ID":"ded25249-eeb5-41f5-b267-32d55af7d2be","Type":"ContainerStarted","Data":"6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c"} Dec 06 11:56:04 crc kubenswrapper[4678]: I1206 11:56:04.789679 4678 generic.go:334] "Generic (PLEG): container finished" podID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerID="6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c" exitCode=0 Dec 06 11:56:04 crc kubenswrapper[4678]: I1206 11:56:04.789728 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcxd" event={"ID":"ded25249-eeb5-41f5-b267-32d55af7d2be","Type":"ContainerDied","Data":"6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c"} Dec 06 11:56:05 crc kubenswrapper[4678]: I1206 11:56:05.799857 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcxd" event={"ID":"ded25249-eeb5-41f5-b267-32d55af7d2be","Type":"ContainerStarted","Data":"a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e"} Dec 06 11:56:05 crc kubenswrapper[4678]: I1206 11:56:05.818480 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xfcxd" podStartSLOduration=2.33839562 podStartE2EDuration="4.81845914s" podCreationTimestamp="2025-12-06 11:56:01 +0000 UTC" firstStartedPulling="2025-12-06 11:56:02.765633079 +0000 UTC m=+4767.609064508" lastFinishedPulling="2025-12-06 11:56:05.245696589 +0000 UTC m=+4770.089128028" observedRunningTime="2025-12-06 11:56:05.813988191 +0000 UTC m=+4770.657419630" watchObservedRunningTime="2025-12-06 11:56:05.81845914 +0000 UTC m=+4770.661890589" Dec 06 11:56:11 crc kubenswrapper[4678]: I1206 11:56:11.377974 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:11 crc kubenswrapper[4678]: I1206 11:56:11.378376 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:11 crc kubenswrapper[4678]: I1206 11:56:11.425406 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:11 crc kubenswrapper[4678]: I1206 11:56:11.908300 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:11 crc kubenswrapper[4678]: I1206 11:56:11.967584 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfcxd"] Dec 06 11:56:13 crc kubenswrapper[4678]: I1206 11:56:13.870594 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xfcxd" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="registry-server" containerID="cri-o://a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e" gracePeriod=2 Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.384175 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.551432 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-utilities\") pod \"ded25249-eeb5-41f5-b267-32d55af7d2be\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.551545 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-catalog-content\") pod \"ded25249-eeb5-41f5-b267-32d55af7d2be\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.551602 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt5q6\" (UniqueName: \"kubernetes.io/projected/ded25249-eeb5-41f5-b267-32d55af7d2be-kube-api-access-zt5q6\") pod \"ded25249-eeb5-41f5-b267-32d55af7d2be\" (UID: \"ded25249-eeb5-41f5-b267-32d55af7d2be\") " Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.553844 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-utilities" (OuterVolumeSpecName: "utilities") pod "ded25249-eeb5-41f5-b267-32d55af7d2be" (UID: "ded25249-eeb5-41f5-b267-32d55af7d2be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.557755 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ded25249-eeb5-41f5-b267-32d55af7d2be-kube-api-access-zt5q6" (OuterVolumeSpecName: "kube-api-access-zt5q6") pod "ded25249-eeb5-41f5-b267-32d55af7d2be" (UID: "ded25249-eeb5-41f5-b267-32d55af7d2be"). InnerVolumeSpecName "kube-api-access-zt5q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.610361 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ded25249-eeb5-41f5-b267-32d55af7d2be" (UID: "ded25249-eeb5-41f5-b267-32d55af7d2be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.654301 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt5q6\" (UniqueName: \"kubernetes.io/projected/ded25249-eeb5-41f5-b267-32d55af7d2be-kube-api-access-zt5q6\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.654344 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.654360 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ded25249-eeb5-41f5-b267-32d55af7d2be-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.881363 4678 generic.go:334] "Generic (PLEG): container finished" podID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerID="a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e" exitCode=0 Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.881420 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcxd" event={"ID":"ded25249-eeb5-41f5-b267-32d55af7d2be","Type":"ContainerDied","Data":"a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e"} Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.881448 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xfcxd" event={"ID":"ded25249-eeb5-41f5-b267-32d55af7d2be","Type":"ContainerDied","Data":"2054400e6cb4045b9eaec6b0e3016819acaf8c6ac9259a420b8bf56ae8bfd5a8"} Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.881480 4678 scope.go:117] "RemoveContainer" containerID="a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.884623 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xfcxd" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.903181 4678 scope.go:117] "RemoveContainer" containerID="6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.941356 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xfcxd"] Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.955810 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xfcxd"] Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.967641 4678 scope.go:117] "RemoveContainer" containerID="814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.998963 4678 scope.go:117] "RemoveContainer" containerID="a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e" Dec 06 11:56:14 crc kubenswrapper[4678]: E1206 11:56:14.999439 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e\": container with ID starting with a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e not found: ID does not exist" containerID="a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.999475 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e"} err="failed to get container status \"a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e\": rpc error: code = NotFound desc = could not find container \"a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e\": container with ID starting with a0d17fe2f473e793b6a6c294b8e1a463e2a882116bfef2192416647ef907a53e not found: ID does not exist" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.999514 4678 scope.go:117] "RemoveContainer" containerID="6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c" Dec 06 11:56:14 crc kubenswrapper[4678]: E1206 11:56:14.999872 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c\": container with ID starting with 6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c not found: ID does not exist" containerID="6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c" Dec 06 11:56:14 crc kubenswrapper[4678]: I1206 11:56:14.999900 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c"} err="failed to get container status \"6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c\": rpc error: code = NotFound desc = could not find container \"6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c\": container with ID starting with 6c1b5802d9eb4a21022526b56a1f99920a4b235564d5733d854c47deb67a727c not found: ID does not exist" Dec 06 11:56:15 crc kubenswrapper[4678]: I1206 11:56:14.999914 4678 scope.go:117] "RemoveContainer" containerID="814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08" Dec 06 11:56:15 crc kubenswrapper[4678]: E1206 11:56:15.000264 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08\": container with ID starting with 814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08 not found: ID does not exist" containerID="814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08" Dec 06 11:56:15 crc kubenswrapper[4678]: I1206 11:56:15.000289 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08"} err="failed to get container status \"814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08\": rpc error: code = NotFound desc = could not find container \"814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08\": container with ID starting with 814331ff375fca4064dacf4c185b601bfa786a9def1c2df2e00f938de8c0ca08 not found: ID does not exist" Dec 06 11:56:15 crc kubenswrapper[4678]: I1206 11:56:15.486409 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" path="/var/lib/kubelet/pods/ded25249-eeb5-41f5-b267-32d55af7d2be/volumes" Dec 06 11:56:30 crc kubenswrapper[4678]: I1206 11:56:30.039701 4678 generic.go:334] "Generic (PLEG): container finished" podID="6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b" containerID="fab1d58958ee3dcb69ef414b57938f17ab2c9213544cb0e498087e94b966ab8e" exitCode=0 Dec 06 11:56:30 crc kubenswrapper[4678]: I1206 11:56:30.039782 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" event={"ID":"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b","Type":"ContainerDied","Data":"fab1d58958ee3dcb69ef414b57938f17ab2c9213544cb0e498087e94b966ab8e"} Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.150814 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.184218 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-shwqz/crc-debug-vtdh2"] Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.192075 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-shwqz/crc-debug-vtdh2"] Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.276019 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwwbh\" (UniqueName: \"kubernetes.io/projected/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-kube-api-access-dwwbh\") pod \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.276145 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-host\") pod \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\" (UID: \"6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b\") " Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.276476 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-host" (OuterVolumeSpecName: "host") pod "6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b" (UID: "6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.280932 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-kube-api-access-dwwbh" (OuterVolumeSpecName: "kube-api-access-dwwbh") pod "6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b" (UID: "6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b"). InnerVolumeSpecName "kube-api-access-dwwbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.378138 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-host\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.378173 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwwbh\" (UniqueName: \"kubernetes.io/projected/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b-kube-api-access-dwwbh\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:31 crc kubenswrapper[4678]: I1206 11:56:31.487751 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b" path="/var/lib/kubelet/pods/6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b/volumes" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.056727 4678 scope.go:117] "RemoveContainer" containerID="fab1d58958ee3dcb69ef414b57938f17ab2c9213544cb0e498087e94b966ab8e" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.057090 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-vtdh2" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.422004 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-shwqz/crc-debug-79gpn"] Dec 06 11:56:32 crc kubenswrapper[4678]: E1206 11:56:32.422362 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="extract-content" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.422373 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="extract-content" Dec 06 11:56:32 crc kubenswrapper[4678]: E1206 11:56:32.422393 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="extract-utilities" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.422399 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="extract-utilities" Dec 06 11:56:32 crc kubenswrapper[4678]: E1206 11:56:32.422420 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b" containerName="container-00" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.422426 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b" containerName="container-00" Dec 06 11:56:32 crc kubenswrapper[4678]: E1206 11:56:32.422445 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="registry-server" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.422466 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="registry-server" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.422644 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e6911ff-f2bb-42d8-9b1b-fd36b2556f0b" containerName="container-00" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.422662 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="ded25249-eeb5-41f5-b267-32d55af7d2be" containerName="registry-server" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.423231 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.497587 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94b24d42-8337-4d2a-83a4-01b8c54486b9-host\") pod \"crc-debug-79gpn\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.497784 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tww9c\" (UniqueName: \"kubernetes.io/projected/94b24d42-8337-4d2a-83a4-01b8c54486b9-kube-api-access-tww9c\") pod \"crc-debug-79gpn\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.599909 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94b24d42-8337-4d2a-83a4-01b8c54486b9-host\") pod \"crc-debug-79gpn\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.600033 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tww9c\" (UniqueName: \"kubernetes.io/projected/94b24d42-8337-4d2a-83a4-01b8c54486b9-kube-api-access-tww9c\") pod \"crc-debug-79gpn\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.600079 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94b24d42-8337-4d2a-83a4-01b8c54486b9-host\") pod \"crc-debug-79gpn\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.617950 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tww9c\" (UniqueName: \"kubernetes.io/projected/94b24d42-8337-4d2a-83a4-01b8c54486b9-kube-api-access-tww9c\") pod \"crc-debug-79gpn\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:32 crc kubenswrapper[4678]: I1206 11:56:32.743945 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:33 crc kubenswrapper[4678]: I1206 11:56:33.066605 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-79gpn" event={"ID":"94b24d42-8337-4d2a-83a4-01b8c54486b9","Type":"ContainerStarted","Data":"97a34c450ae72a01e3def79b1fa7175e3e9b079f667ac2b51b7ffa74f2fbf368"} Dec 06 11:56:33 crc kubenswrapper[4678]: I1206 11:56:33.066658 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-79gpn" event={"ID":"94b24d42-8337-4d2a-83a4-01b8c54486b9","Type":"ContainerStarted","Data":"5a4bd9ccad5341e9ca2dd2f049c5672ce0eade36bbb074a0dcd5471b90b98492"} Dec 06 11:56:33 crc kubenswrapper[4678]: I1206 11:56:33.078745 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-shwqz/crc-debug-79gpn" podStartSLOduration=1.078729244 podStartE2EDuration="1.078729244s" podCreationTimestamp="2025-12-06 11:56:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-06 11:56:33.077832596 +0000 UTC m=+4797.921264035" watchObservedRunningTime="2025-12-06 11:56:33.078729244 +0000 UTC m=+4797.922160683" Dec 06 11:56:34 crc kubenswrapper[4678]: I1206 11:56:34.074960 4678 generic.go:334] "Generic (PLEG): container finished" podID="94b24d42-8337-4d2a-83a4-01b8c54486b9" containerID="97a34c450ae72a01e3def79b1fa7175e3e9b079f667ac2b51b7ffa74f2fbf368" exitCode=0 Dec 06 11:56:34 crc kubenswrapper[4678]: I1206 11:56:34.074986 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-79gpn" event={"ID":"94b24d42-8337-4d2a-83a4-01b8c54486b9","Type":"ContainerDied","Data":"97a34c450ae72a01e3def79b1fa7175e3e9b079f667ac2b51b7ffa74f2fbf368"} Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.176562 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.340917 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tww9c\" (UniqueName: \"kubernetes.io/projected/94b24d42-8337-4d2a-83a4-01b8c54486b9-kube-api-access-tww9c\") pod \"94b24d42-8337-4d2a-83a4-01b8c54486b9\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.341281 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94b24d42-8337-4d2a-83a4-01b8c54486b9-host\") pod \"94b24d42-8337-4d2a-83a4-01b8c54486b9\" (UID: \"94b24d42-8337-4d2a-83a4-01b8c54486b9\") " Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.341685 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/94b24d42-8337-4d2a-83a4-01b8c54486b9-host" (OuterVolumeSpecName: "host") pod "94b24d42-8337-4d2a-83a4-01b8c54486b9" (UID: "94b24d42-8337-4d2a-83a4-01b8c54486b9"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.341933 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/94b24d42-8337-4d2a-83a4-01b8c54486b9-host\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.346076 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94b24d42-8337-4d2a-83a4-01b8c54486b9-kube-api-access-tww9c" (OuterVolumeSpecName: "kube-api-access-tww9c") pod "94b24d42-8337-4d2a-83a4-01b8c54486b9" (UID: "94b24d42-8337-4d2a-83a4-01b8c54486b9"). InnerVolumeSpecName "kube-api-access-tww9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.371334 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-shwqz/crc-debug-79gpn"] Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.384903 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-shwqz/crc-debug-79gpn"] Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.443619 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tww9c\" (UniqueName: \"kubernetes.io/projected/94b24d42-8337-4d2a-83a4-01b8c54486b9-kube-api-access-tww9c\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:35 crc kubenswrapper[4678]: I1206 11:56:35.536881 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94b24d42-8337-4d2a-83a4-01b8c54486b9" path="/var/lib/kubelet/pods/94b24d42-8337-4d2a-83a4-01b8c54486b9/volumes" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.093269 4678 scope.go:117] "RemoveContainer" containerID="97a34c450ae72a01e3def79b1fa7175e3e9b079f667ac2b51b7ffa74f2fbf368" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.093344 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-79gpn" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.678470 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-shwqz/crc-debug-2qwps"] Dec 06 11:56:36 crc kubenswrapper[4678]: E1206 11:56:36.679179 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94b24d42-8337-4d2a-83a4-01b8c54486b9" containerName="container-00" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.679193 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="94b24d42-8337-4d2a-83a4-01b8c54486b9" containerName="container-00" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.679382 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="94b24d42-8337-4d2a-83a4-01b8c54486b9" containerName="container-00" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.680019 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.867692 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940f47d0-2037-441a-b80f-b94c1d638363-host\") pod \"crc-debug-2qwps\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.868157 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fk9r\" (UniqueName: \"kubernetes.io/projected/940f47d0-2037-441a-b80f-b94c1d638363-kube-api-access-9fk9r\") pod \"crc-debug-2qwps\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.970564 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940f47d0-2037-441a-b80f-b94c1d638363-host\") pod \"crc-debug-2qwps\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.970682 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fk9r\" (UniqueName: \"kubernetes.io/projected/940f47d0-2037-441a-b80f-b94c1d638363-kube-api-access-9fk9r\") pod \"crc-debug-2qwps\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:36 crc kubenswrapper[4678]: I1206 11:56:36.970729 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940f47d0-2037-441a-b80f-b94c1d638363-host\") pod \"crc-debug-2qwps\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:37 crc kubenswrapper[4678]: I1206 11:56:37.002100 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fk9r\" (UniqueName: \"kubernetes.io/projected/940f47d0-2037-441a-b80f-b94c1d638363-kube-api-access-9fk9r\") pod \"crc-debug-2qwps\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:37 crc kubenswrapper[4678]: I1206 11:56:37.295866 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:38 crc kubenswrapper[4678]: I1206 11:56:38.113865 4678 generic.go:334] "Generic (PLEG): container finished" podID="940f47d0-2037-441a-b80f-b94c1d638363" containerID="0bd6b5dbbbb47589ea845db3051fdc960852dca237d81595ee7fb957dbd870cb" exitCode=0 Dec 06 11:56:38 crc kubenswrapper[4678]: I1206 11:56:38.113956 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-2qwps" event={"ID":"940f47d0-2037-441a-b80f-b94c1d638363","Type":"ContainerDied","Data":"0bd6b5dbbbb47589ea845db3051fdc960852dca237d81595ee7fb957dbd870cb"} Dec 06 11:56:38 crc kubenswrapper[4678]: I1206 11:56:38.114257 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/crc-debug-2qwps" event={"ID":"940f47d0-2037-441a-b80f-b94c1d638363","Type":"ContainerStarted","Data":"b0a20b32fca4f0d9aed77179beedb3145f3514f431d9731054fe7658bbe2b790"} Dec 06 11:56:38 crc kubenswrapper[4678]: I1206 11:56:38.149867 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-shwqz/crc-debug-2qwps"] Dec 06 11:56:38 crc kubenswrapper[4678]: I1206 11:56:38.160091 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-shwqz/crc-debug-2qwps"] Dec 06 11:56:39 crc kubenswrapper[4678]: I1206 11:56:39.582659 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:39 crc kubenswrapper[4678]: I1206 11:56:39.717260 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fk9r\" (UniqueName: \"kubernetes.io/projected/940f47d0-2037-441a-b80f-b94c1d638363-kube-api-access-9fk9r\") pod \"940f47d0-2037-441a-b80f-b94c1d638363\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " Dec 06 11:56:39 crc kubenswrapper[4678]: I1206 11:56:39.717404 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940f47d0-2037-441a-b80f-b94c1d638363-host\") pod \"940f47d0-2037-441a-b80f-b94c1d638363\" (UID: \"940f47d0-2037-441a-b80f-b94c1d638363\") " Dec 06 11:56:39 crc kubenswrapper[4678]: I1206 11:56:39.717694 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/940f47d0-2037-441a-b80f-b94c1d638363-host" (OuterVolumeSpecName: "host") pod "940f47d0-2037-441a-b80f-b94c1d638363" (UID: "940f47d0-2037-441a-b80f-b94c1d638363"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 06 11:56:39 crc kubenswrapper[4678]: I1206 11:56:39.718244 4678 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/940f47d0-2037-441a-b80f-b94c1d638363-host\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:39 crc kubenswrapper[4678]: I1206 11:56:39.727163 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/940f47d0-2037-441a-b80f-b94c1d638363-kube-api-access-9fk9r" (OuterVolumeSpecName: "kube-api-access-9fk9r") pod "940f47d0-2037-441a-b80f-b94c1d638363" (UID: "940f47d0-2037-441a-b80f-b94c1d638363"). InnerVolumeSpecName "kube-api-access-9fk9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 11:56:39 crc kubenswrapper[4678]: I1206 11:56:39.820264 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fk9r\" (UniqueName: \"kubernetes.io/projected/940f47d0-2037-441a-b80f-b94c1d638363-kube-api-access-9fk9r\") on node \"crc\" DevicePath \"\"" Dec 06 11:56:40 crc kubenswrapper[4678]: I1206 11:56:40.132910 4678 scope.go:117] "RemoveContainer" containerID="0bd6b5dbbbb47589ea845db3051fdc960852dca237d81595ee7fb957dbd870cb" Dec 06 11:56:40 crc kubenswrapper[4678]: I1206 11:56:40.133083 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/crc-debug-2qwps" Dec 06 11:56:41 crc kubenswrapper[4678]: E1206 11:56:41.327182 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b24d42_8337_4d2a_83a4_01b8c54486b9.slice\": RecentStats: unable to find data in memory cache]" Dec 06 11:56:41 crc kubenswrapper[4678]: I1206 11:56:41.485641 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="940f47d0-2037-441a-b80f-b94c1d638363" path="/var/lib/kubelet/pods/940f47d0-2037-441a-b80f-b94c1d638363/volumes" Dec 06 11:56:51 crc kubenswrapper[4678]: E1206 11:56:51.653364 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b24d42_8337_4d2a_83a4_01b8c54486b9.slice\": RecentStats: unable to find data in memory cache]" Dec 06 11:56:59 crc kubenswrapper[4678]: I1206 11:56:59.505502 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:56:59 crc kubenswrapper[4678]: I1206 11:56:59.507007 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:57:01 crc kubenswrapper[4678]: E1206 11:57:01.944578 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b24d42_8337_4d2a_83a4_01b8c54486b9.slice\": RecentStats: unable to find data in memory cache]" Dec 06 11:57:12 crc kubenswrapper[4678]: E1206 11:57:12.176023 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b24d42_8337_4d2a_83a4_01b8c54486b9.slice\": RecentStats: unable to find data in memory cache]" Dec 06 11:57:16 crc kubenswrapper[4678]: I1206 11:57:16.732383 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7b874f6c6d-jcklm_6e0b0cc3-b79e-4e3a-8459-adba09620269/barbican-api/0.log" Dec 06 11:57:16 crc kubenswrapper[4678]: I1206 11:57:16.746265 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7b874f6c6d-jcklm_6e0b0cc3-b79e-4e3a-8459-adba09620269/barbican-api-log/0.log" Dec 06 11:57:16 crc kubenswrapper[4678]: I1206 11:57:16.980097 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5674df646-s4z6g_e55dbe47-fb3b-4020-bba4-f8c038e7cc5b/barbican-keystone-listener/0.log" Dec 06 11:57:17 crc kubenswrapper[4678]: I1206 11:57:17.021743 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5674df646-s4z6g_e55dbe47-fb3b-4020-bba4-f8c038e7cc5b/barbican-keystone-listener-log/0.log" Dec 06 11:57:17 crc kubenswrapper[4678]: I1206 11:57:17.146105 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-597bfcdc9c-vdzb7_69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957/barbican-worker/0.log" Dec 06 11:57:17 crc kubenswrapper[4678]: I1206 11:57:17.217642 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-597bfcdc9c-vdzb7_69b9a5eb-f2d2-423e-b3c6-1f4b5b8ef957/barbican-worker-log/0.log" Dec 06 11:57:17 crc kubenswrapper[4678]: I1206 11:57:17.375793 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-59bpf_99d806d5-73ca-4ecc-92cb-b3c300cb04b4/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:17 crc kubenswrapper[4678]: I1206 11:57:17.520160 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/ceilometer-central-agent/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.142996 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/ceilometer-notification-agent/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.161365 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/proxy-httpd/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.180641 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_5a9c85b9-5f97-4449-ba76-0f81fd00cce1/sg-core/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.417256 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b283a032-2cad-40a7-bb76-88a4f3f47472/cinder-api-log/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.501305 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b283a032-2cad-40a7-bb76-88a4f3f47472/cinder-api/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.649831 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_918bbdd7-35bb-457d-a29f-5c6e4f04b062/cinder-scheduler/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.771278 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_918bbdd7-35bb-457d-a29f-5c6e4f04b062/probe/0.log" Dec 06 11:57:18 crc kubenswrapper[4678]: I1206 11:57:18.996222 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-hrxqj_5fa31310-7e0f-431a-ae13-287fd25a926c/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:19 crc kubenswrapper[4678]: I1206 11:57:19.127929 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-k2hf8_f104b9bb-12fa-4f5c-86ed-2db61119cb55/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:19 crc kubenswrapper[4678]: I1206 11:57:19.316961 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67cb876dc9-5749h_1ab593af-6400-4c78-b798-fd34038f28dd/init/0.log" Dec 06 11:57:19 crc kubenswrapper[4678]: I1206 11:57:19.553020 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67cb876dc9-5749h_1ab593af-6400-4c78-b798-fd34038f28dd/init/0.log" Dec 06 11:57:19 crc kubenswrapper[4678]: I1206 11:57:19.788099 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67cb876dc9-5749h_1ab593af-6400-4c78-b798-fd34038f28dd/dnsmasq-dns/0.log" Dec 06 11:57:19 crc kubenswrapper[4678]: I1206 11:57:19.807151 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-877bb_9ac6052e-a10a-4709-b1be-99bb37d0922d/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:20 crc kubenswrapper[4678]: I1206 11:57:20.171799 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ef3105b5-af55-4aec-99a8-10792740d742/glance-httpd/0.log" Dec 06 11:57:20 crc kubenswrapper[4678]: I1206 11:57:20.274952 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_ef3105b5-af55-4aec-99a8-10792740d742/glance-log/0.log" Dec 06 11:57:20 crc kubenswrapper[4678]: I1206 11:57:20.504542 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_965cdf20-a198-4cc8-9a8f-700388dbf4cd/glance-log/0.log" Dec 06 11:57:20 crc kubenswrapper[4678]: I1206 11:57:20.535638 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_965cdf20-a198-4cc8-9a8f-700388dbf4cd/glance-httpd/0.log" Dec 06 11:57:20 crc kubenswrapper[4678]: I1206 11:57:20.674374 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d77bcbbf8-cfc4l_c2ceb969-bd57-4345-840c-4cd8b2b7ca8e/horizon/1.log" Dec 06 11:57:20 crc kubenswrapper[4678]: I1206 11:57:20.900588 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d77bcbbf8-cfc4l_c2ceb969-bd57-4345-840c-4cd8b2b7ca8e/horizon/0.log" Dec 06 11:57:21 crc kubenswrapper[4678]: I1206 11:57:21.092920 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-fzrzt_a8f9dc0d-5e17-4e1b-9363-2d193c592e9b/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:21 crc kubenswrapper[4678]: I1206 11:57:21.180377 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-7d77bcbbf8-cfc4l_c2ceb969-bd57-4345-840c-4cd8b2b7ca8e/horizon-log/0.log" Dec 06 11:57:21 crc kubenswrapper[4678]: I1206 11:57:21.246682 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4l22m_8e6b45b8-d8db-457e-9a08-04ce67b197f6/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:21 crc kubenswrapper[4678]: I1206 11:57:21.394635 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29416981-627fr_29f472e5-54ad-470c-a43b-a6fbee7fa24d/keystone-cron/0.log" Dec 06 11:57:21 crc kubenswrapper[4678]: I1206 11:57:21.596321 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_8c13e227-c6eb-46a3-94a1-e9a3f3d7a6c3/kube-state-metrics/0.log" Dec 06 11:57:21 crc kubenswrapper[4678]: I1206 11:57:21.841298 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7f69d54bd9-rd99t_eb641828-16dd-4437-8266-e20a7a27ad1c/keystone-api/0.log" Dec 06 11:57:21 crc kubenswrapper[4678]: I1206 11:57:21.868854 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-4lqfl_ff8f7b93-13e6-49a3-937a-8247a3eaf5c8/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:22 crc kubenswrapper[4678]: I1206 11:57:22.444564 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-z49v6_1e7abea2-75f9-4bcb-897d-d62fe1e7be3e/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:22 crc kubenswrapper[4678]: E1206 11:57:22.511372 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b24d42_8337_4d2a_83a4_01b8c54486b9.slice\": RecentStats: unable to find data in memory cache]" Dec 06 11:57:22 crc kubenswrapper[4678]: I1206 11:57:22.654399 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7567449455-km9cw_0e386ecc-df37-4e08-a9f9-80abb7019c19/neutron-httpd/0.log" Dec 06 11:57:23 crc kubenswrapper[4678]: I1206 11:57:23.114576 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-7567449455-km9cw_0e386ecc-df37-4e08-a9f9-80abb7019c19/neutron-api/0.log" Dec 06 11:57:23 crc kubenswrapper[4678]: I1206 11:57:23.892405 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_a92133d6-4bca-447f-9b4c-a1e776a4cb40/nova-cell0-conductor-conductor/0.log" Dec 06 11:57:24 crc kubenswrapper[4678]: I1206 11:57:24.067780 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_ed1b2724-cbb5-4589-9622-6ff20ab1f180/nova-cell1-conductor-conductor/0.log" Dec 06 11:57:24 crc kubenswrapper[4678]: I1206 11:57:24.071014 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4859acf4-20ed-4af8-a83d-9c6303eec9d4/nova-api-log/0.log" Dec 06 11:57:24 crc kubenswrapper[4678]: I1206 11:57:24.504338 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4859acf4-20ed-4af8-a83d-9c6303eec9d4/nova-api-api/0.log" Dec 06 11:57:24 crc kubenswrapper[4678]: I1206 11:57:24.552049 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_21d64141-1313-4582-ac36-a0d0a304f0fb/nova-cell1-novncproxy-novncproxy/0.log" Dec 06 11:57:24 crc kubenswrapper[4678]: I1206 11:57:24.610081 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-4fnld_ff53c201-513d-4b60-b0ef-a1a59a30bd4f/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:25 crc kubenswrapper[4678]: I1206 11:57:25.648106 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_63c277fe-de03-4243-98a2-8d7efaa92c4a/nova-metadata-log/0.log" Dec 06 11:57:25 crc kubenswrapper[4678]: I1206 11:57:25.884769 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1cb14647-bb80-422d-8f59-985a70bf14c3/mysql-bootstrap/0.log" Dec 06 11:57:26 crc kubenswrapper[4678]: I1206 11:57:26.030895 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_a2fb742d-13dd-468a-918e-37c53f81b294/nova-scheduler-scheduler/0.log" Dec 06 11:57:26 crc kubenswrapper[4678]: I1206 11:57:26.166327 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1cb14647-bb80-422d-8f59-985a70bf14c3/mysql-bootstrap/0.log" Dec 06 11:57:26 crc kubenswrapper[4678]: I1206 11:57:26.173454 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1cb14647-bb80-422d-8f59-985a70bf14c3/galera/0.log" Dec 06 11:57:26 crc kubenswrapper[4678]: I1206 11:57:26.759164 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_747aae14-74af-41d1-af19-610b0fa68be1/mysql-bootstrap/0.log" Dec 06 11:57:26 crc kubenswrapper[4678]: I1206 11:57:26.962428 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_747aae14-74af-41d1-af19-610b0fa68be1/galera/0.log" Dec 06 11:57:26 crc kubenswrapper[4678]: I1206 11:57:26.968248 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_747aae14-74af-41d1-af19-610b0fa68be1/mysql-bootstrap/0.log" Dec 06 11:57:27 crc kubenswrapper[4678]: I1206 11:57:27.385219 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_63c277fe-de03-4243-98a2-8d7efaa92c4a/nova-metadata-metadata/0.log" Dec 06 11:57:27 crc kubenswrapper[4678]: I1206 11:57:27.579038 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_66756fbd-c49f-4ea6-8595-7801a7ddecf3/openstackclient/0.log" Dec 06 11:57:27 crc kubenswrapper[4678]: I1206 11:57:27.662598 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-kdb5q_c2658168-1355-4595-bf23-570a68c90da1/ovn-controller/0.log" Dec 06 11:57:27 crc kubenswrapper[4678]: I1206 11:57:27.797224 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-dgln2_f7dd8a89-6c2f-4dcf-96ed-1e9a04bdec69/openstack-network-exporter/0.log" Dec 06 11:57:27 crc kubenswrapper[4678]: I1206 11:57:27.971663 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovsdb-server-init/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.189366 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovsdb-server-init/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.263433 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovs-vswitchd/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.283852 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zcg6x_0155b73c-73e9-4b08-830d-f78f260d6ae5/ovsdb-server/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.560252 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-7xqv9_c79d4229-477e-4f22-90d8-449984bab4cb/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.677619 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5a586aa5-fe85-49da-a10e-3e29de7dba4d/openstack-network-exporter/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.818651 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_5a586aa5-fe85-49da-a10e-3e29de7dba4d/ovn-northd/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.914966 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12052ef6-7a13-4ad5-9484-bbad2cabdf5a/openstack-network-exporter/0.log" Dec 06 11:57:28 crc kubenswrapper[4678]: I1206 11:57:28.957922 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_12052ef6-7a13-4ad5-9484-bbad2cabdf5a/ovsdbserver-nb/0.log" Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.251276 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3f6c33f7-7b74-437b-a9fd-3790af864af7/openstack-network-exporter/0.log" Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.305701 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_3f6c33f7-7b74-437b-a9fd-3790af864af7/ovsdbserver-sb/0.log" Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.505153 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.505205 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.611597 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3/setup-container/0.log" Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.667984 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bd747c594-kkr56_e209518e-5689-468e-9b70-9c1e2864eb9b/placement-api/0.log" Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.833936 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5bd747c594-kkr56_e209518e-5689-468e-9b70-9c1e2864eb9b/placement-log/0.log" Dec 06 11:57:29 crc kubenswrapper[4678]: I1206 11:57:29.886916 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3/setup-container/0.log" Dec 06 11:57:30 crc kubenswrapper[4678]: I1206 11:57:30.000660 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_f7fe7aa1-62b5-4c6b-be2e-b64fce2711b3/rabbitmq/0.log" Dec 06 11:57:30 crc kubenswrapper[4678]: I1206 11:57:30.139127 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cb4272e8-bb6c-43d4-8539-baeb4f3c14a8/setup-container/0.log" Dec 06 11:57:30 crc kubenswrapper[4678]: I1206 11:57:30.440839 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cb4272e8-bb6c-43d4-8539-baeb4f3c14a8/setup-container/0.log" Dec 06 11:57:30 crc kubenswrapper[4678]: I1206 11:57:30.565778 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_cb4272e8-bb6c-43d4-8539-baeb4f3c14a8/rabbitmq/0.log" Dec 06 11:57:30 crc kubenswrapper[4678]: I1206 11:57:30.593304 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-n5rlz_d6e6e2e2-4745-46f6-a3fa-94f803ac0677/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:30 crc kubenswrapper[4678]: I1206 11:57:30.758251 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-jb8nh_d66737d1-caae-4204-9786-2b19a88f5f82/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:31 crc kubenswrapper[4678]: I1206 11:57:31.201680 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-djgvb_29b633cb-d39d-43b9-9556-461cd563c92f/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:31 crc kubenswrapper[4678]: I1206 11:57:31.390831 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-7kqwq_fdc4c25f-dc36-4a52-a26d-9dd390c23061/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:31 crc kubenswrapper[4678]: I1206 11:57:31.463832 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-9pvlf_06c1a00f-a8f0-4ed8-8307-a9ad3d50c0ed/ssh-known-hosts-edpm-deployment/0.log" Dec 06 11:57:31 crc kubenswrapper[4678]: I1206 11:57:31.723609 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f5f494679-7bm9s_b31f2128-db6e-498b-bd38-d0b1c41a4603/proxy-server/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.003957 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-tkg52_db2561ff-632a-463d-a7af-79d83536812b/swift-ring-rebalance/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.018172 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-f5f494679-7bm9s_b31f2128-db6e-498b-bd38-d0b1c41a4603/proxy-httpd/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.053829 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-auditor/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.308651 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-reaper/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.311258 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-replicator/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.377846 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/account-server/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.454097 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-auditor/0.log" Dec 06 11:57:32 crc kubenswrapper[4678]: E1206 11:57:32.814939 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94b24d42_8337_4d2a_83a4_01b8c54486b9.slice\": RecentStats: unable to find data in memory cache]" Dec 06 11:57:32 crc kubenswrapper[4678]: I1206 11:57:32.998810 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-server/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.054282 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-replicator/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.087544 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/container-updater/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.142244 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-auditor/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.211276 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-expirer/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.314898 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-replicator/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.385229 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-server/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.440867 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/object-updater/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.539400 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/rsync/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.631991 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_a7f4ca97-1480-4327-8025-9b18a8a7073b/swift-recon-cron/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.804156 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-dj678_1cadf619-4b68-44a3-a65b-422bea0a2ede/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:33 crc kubenswrapper[4678]: I1206 11:57:33.989973 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_5b32697e-d8b7-4852-8afe-a6f19920f210/tempest-tests-tempest-tests-runner/0.log" Dec 06 11:57:34 crc kubenswrapper[4678]: I1206 11:57:34.489241 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_3b09e665-d6c6-4ac4-831b-210905b8fa20/test-operator-logs-container/0.log" Dec 06 11:57:34 crc kubenswrapper[4678]: I1206 11:57:34.880726 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-5xnzs_89741e1b-d65e-4e23-982d-716c2e498c25/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 06 11:57:46 crc kubenswrapper[4678]: I1206 11:57:46.194395 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_a0c958d8-87e6-4b93-be23-58cb84d32dd4/memcached/0.log" Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.505072 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.505595 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.505633 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.506271 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a8c3c8b241bd53b83426412a64cbd2cddf1cbe5ae5bf0d01ab9debe1412cf227"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.506314 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://a8c3c8b241bd53b83426412a64cbd2cddf1cbe5ae5bf0d01ab9debe1412cf227" gracePeriod=600 Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.880620 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="a8c3c8b241bd53b83426412a64cbd2cddf1cbe5ae5bf0d01ab9debe1412cf227" exitCode=0 Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.880720 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"a8c3c8b241bd53b83426412a64cbd2cddf1cbe5ae5bf0d01ab9debe1412cf227"} Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.880952 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412"} Dec 06 11:57:59 crc kubenswrapper[4678]: I1206 11:57:59.880976 4678 scope.go:117] "RemoveContainer" containerID="c8619ab527cbc04446cbc5cfe73cc8c96028c9251185adf7790039d437a35628" Dec 06 11:58:08 crc kubenswrapper[4678]: I1206 11:58:08.305941 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/util/0.log" Dec 06 11:58:08 crc kubenswrapper[4678]: I1206 11:58:08.992277 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/util/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.037144 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/pull/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.056881 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/pull/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.222125 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/util/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.230240 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/pull/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.273183 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_67f979f2024db09399f08cb4db17df395f8f76770ad48fa29cd9e772d07kctd_e517ee42-f060-4681-97fd-386010a0fc4d/extract/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.441606 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-th2qw_6669b3a4-9c79-441e-90a2-abceaa89fc58/kube-rbac-proxy/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.602235 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-th2qw_6669b3a4-9c79-441e-90a2-abceaa89fc58/manager/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.636820 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zftrw_e48c5af7-e126-49e1-9bc0-eec4bd474080/kube-rbac-proxy/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.696719 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-zftrw_e48c5af7-e126-49e1-9bc0-eec4bd474080/manager/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.866690 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-xqpcb_807c6a8a-a12e-4ac6-8d6a-01f54876d4dd/manager/0.log" Dec 06 11:58:09 crc kubenswrapper[4678]: I1206 11:58:09.885008 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-xqpcb_807c6a8a-a12e-4ac6-8d6a-01f54876d4dd/kube-rbac-proxy/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.045873 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-jpgnt_961960ba-f29c-40e1-81c4-5f8b43456dca/kube-rbac-proxy/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.166037 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-jpgnt_961960ba-f29c-40e1-81c4-5f8b43456dca/manager/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.199056 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-ktmz7_c58c8710-6867-43d2-8fb4-ea96041e2422/kube-rbac-proxy/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.313938 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-ktmz7_c58c8710-6867-43d2-8fb4-ea96041e2422/manager/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.438034 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9kgbd_b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef/kube-rbac-proxy/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.467238 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-9kgbd_b1c49eb6-f184-4fc1-83d2-e3b86c2a19ef/manager/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.639276 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-9svrt_d272c78c-4c56-4faa-a421-ede41c4b2307/kube-rbac-proxy/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.905079 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-9svrt_d272c78c-4c56-4faa-a421-ede41c4b2307/manager/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.926471 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-ssssw_7bc20ad2-7338-4943-ae21-1ad8a0fd947e/kube-rbac-proxy/0.log" Dec 06 11:58:10 crc kubenswrapper[4678]: I1206 11:58:10.937912 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-ssssw_7bc20ad2-7338-4943-ae21-1ad8a0fd947e/manager/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.178184 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p5bxv_78fc5203-6af1-49e0-a485-ba9ef0a0b658/manager/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.184870 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-p5bxv_78fc5203-6af1-49e0-a485-ba9ef0a0b658/kube-rbac-proxy/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.223369 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-vdspj_9e2b27dd-544c-4c0a-ba0f-6c0b4243017c/kube-rbac-proxy/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.371189 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-vdspj_9e2b27dd-544c-4c0a-ba0f-6c0b4243017c/manager/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.446503 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9b7t6_c7983f9e-7a85-4c2c-b689-7213f06a144a/kube-rbac-proxy/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.520792 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9b7t6_c7983f9e-7a85-4c2c-b689-7213f06a144a/manager/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.619180 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-d5h8j_d925b372-ccef-42a2-9b2c-de8539f9d070/kube-rbac-proxy/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.734924 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-d5h8j_d925b372-ccef-42a2-9b2c-de8539f9d070/manager/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.871701 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4kvsf_6424c424-5901-4d42-a077-8388ef0978b8/kube-rbac-proxy/0.log" Dec 06 11:58:11 crc kubenswrapper[4678]: I1206 11:58:11.980888 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-4kvsf_6424c424-5901-4d42-a077-8388ef0978b8/manager/0.log" Dec 06 11:58:12 crc kubenswrapper[4678]: I1206 11:58:12.020313 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fpj6x_f1b9824d-0c1c-4cef-bb30-3e0f14719c6d/kube-rbac-proxy/0.log" Dec 06 11:58:12 crc kubenswrapper[4678]: I1206 11:58:12.088733 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-fpj6x_f1b9824d-0c1c-4cef-bb30-3e0f14719c6d/manager/0.log" Dec 06 11:58:12 crc kubenswrapper[4678]: I1206 11:58:12.208741 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fs2699_20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f/kube-rbac-proxy/0.log" Dec 06 11:58:12 crc kubenswrapper[4678]: I1206 11:58:12.211882 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-84b575879fs2699_20f3fdfd-d2c8-427c-8371-e5d9cefe0c7f/manager/0.log" Dec 06 11:58:12 crc kubenswrapper[4678]: I1206 11:58:12.600378 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5bbffdd548-5zgrs_0b336e20-2347-4b86-86b4-22ef7da8983c/operator/0.log" Dec 06 11:58:12 crc kubenswrapper[4678]: I1206 11:58:12.801836 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-m7422_dfb6011a-3444-4e31-8a84-d2c521f55e77/registry-server/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.006858 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mrqtn_ed890ee4-e09b-40cd-9bce-f21d8676a2c7/kube-rbac-proxy/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.206188 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-mrqtn_ed890ee4-e09b-40cd-9bce-f21d8676a2c7/manager/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.323670 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-jdg4h_f1447b71-f9c7-4969-9d8a-d4b8b1f099bb/kube-rbac-proxy/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.326371 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-jdg4h_f1447b71-f9c7-4969-9d8a-d4b8b1f099bb/manager/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.483317 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-wtgcf_c232b1db-4726-4697-b59f-f80c1d9d2d4d/operator/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.612863 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5dcc4c7f94-xld2p_67b34972-6c66-47d8-885e-38d2a55a4fd9/manager/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.628519 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-pknvj_b0e3bc2c-84eb-4407-ae9b-101a70b9d555/kube-rbac-proxy/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.754567 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-pknvj_b0e3bc2c-84eb-4407-ae9b-101a70b9d555/manager/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.805282 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-m4mrc_b458a451-d4c7-4cb3-b7af-90a12424be9b/kube-rbac-proxy/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.947986 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-m4mrc_b458a451-d4c7-4cb3-b7af-90a12424be9b/manager/0.log" Dec 06 11:58:13 crc kubenswrapper[4678]: I1206 11:58:13.990915 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qj7fr_9b1bcff8-3a58-4c8a-a971-0374ecb16a9b/kube-rbac-proxy/0.log" Dec 06 11:58:14 crc kubenswrapper[4678]: I1206 11:58:14.039863 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-qj7fr_9b1bcff8-3a58-4c8a-a971-0374ecb16a9b/manager/0.log" Dec 06 11:58:14 crc kubenswrapper[4678]: I1206 11:58:14.142945 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-wl7ww_90b306bd-29a2-494d-8399-75d95e9b8597/kube-rbac-proxy/0.log" Dec 06 11:58:14 crc kubenswrapper[4678]: I1206 11:58:14.169034 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-wl7ww_90b306bd-29a2-494d-8399-75d95e9b8597/manager/0.log" Dec 06 11:58:34 crc kubenswrapper[4678]: I1206 11:58:34.926430 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dvssp_9b0941dc-500a-4284-9cc4-66b250feee93/control-plane-machine-set-operator/0.log" Dec 06 11:58:35 crc kubenswrapper[4678]: I1206 11:58:35.223037 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r5rng_0061f615-2fb9-4e17-8ba5-62d97e8070c8/kube-rbac-proxy/0.log" Dec 06 11:58:35 crc kubenswrapper[4678]: I1206 11:58:35.282644 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-r5rng_0061f615-2fb9-4e17-8ba5-62d97e8070c8/machine-api-operator/0.log" Dec 06 11:58:50 crc kubenswrapper[4678]: I1206 11:58:50.748942 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-ckqhf_048fff94-2a2e-4819-af68-f3e8b2f2fa3f/cert-manager-controller/0.log" Dec 06 11:58:50 crc kubenswrapper[4678]: I1206 11:58:50.790117 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-8889z_882b9c0d-2380-42b7-b13f-480a4f53a258/cert-manager-cainjector/0.log" Dec 06 11:58:50 crc kubenswrapper[4678]: I1206 11:58:50.959156 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-rhwnp_5e3d90b8-1cbb-4329-8d55-911e6dad3243/cert-manager-webhook/0.log" Dec 06 11:59:03 crc kubenswrapper[4678]: I1206 11:59:03.840651 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-g72ds_d99e972f-a316-4107-85b4-bf5411999e91/nmstate-console-plugin/0.log" Dec 06 11:59:04 crc kubenswrapper[4678]: I1206 11:59:04.058977 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-n2j4m_82398c4a-8182-4723-bad7-0c289be9d283/nmstate-handler/0.log" Dec 06 11:59:04 crc kubenswrapper[4678]: I1206 11:59:04.127311 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-7nc8k_1e102849-cdd0-4035-a9ba-339d77bd6f8e/nmstate-metrics/0.log" Dec 06 11:59:04 crc kubenswrapper[4678]: I1206 11:59:04.137663 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-7nc8k_1e102849-cdd0-4035-a9ba-339d77bd6f8e/kube-rbac-proxy/0.log" Dec 06 11:59:04 crc kubenswrapper[4678]: I1206 11:59:04.374553 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-5rqmh_fd2552d3-a5fd-4b13-8d01-996a25348826/nmstate-operator/0.log" Dec 06 11:59:04 crc kubenswrapper[4678]: I1206 11:59:04.399354 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-8gwjz_ca2c68b1-3733-4be8-9580-616e9d436a34/nmstate-webhook/0.log" Dec 06 11:59:20 crc kubenswrapper[4678]: I1206 11:59:20.803117 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-wstfq_714f1cf8-6f0e-4c0f-8924-4b07a98c578f/kube-rbac-proxy/0.log" Dec 06 11:59:20 crc kubenswrapper[4678]: I1206 11:59:20.859926 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-wstfq_714f1cf8-6f0e-4c0f-8924-4b07a98c578f/controller/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.031331 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.211070 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.264679 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.287289 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.318300 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.491243 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.545925 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.562292 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.636986 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.772932 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-frr-files/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.801288 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-reloader/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.809126 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/cp-metrics/0.log" Dec 06 11:59:21 crc kubenswrapper[4678]: I1206 11:59:21.859871 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/controller/0.log" Dec 06 11:59:22 crc kubenswrapper[4678]: I1206 11:59:22.036785 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/kube-rbac-proxy/0.log" Dec 06 11:59:22 crc kubenswrapper[4678]: I1206 11:59:22.053630 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/frr-metrics/0.log" Dec 06 11:59:22 crc kubenswrapper[4678]: I1206 11:59:22.254212 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/kube-rbac-proxy-frr/0.log" Dec 06 11:59:22 crc kubenswrapper[4678]: I1206 11:59:22.337532 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/reloader/0.log" Dec 06 11:59:22 crc kubenswrapper[4678]: I1206 11:59:22.511954 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-sqk7m_1582c836-cc2c-4d01-b878-6fd5cde0ae70/frr-k8s-webhook-server/0.log" Dec 06 11:59:22 crc kubenswrapper[4678]: I1206 11:59:22.742744 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-dd988988c-fc549_bc881548-41a9-4790-b691-dfb5935d128f/manager/0.log" Dec 06 11:59:22 crc kubenswrapper[4678]: I1206 11:59:22.931704 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d6f755774-dghr8_2562672a-fcdf-420a-a4df-116113c240a0/webhook-server/0.log" Dec 06 11:59:23 crc kubenswrapper[4678]: I1206 11:59:23.118135 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qjlkq_e6542baf-28b2-4844-991e-eb0f57d45231/kube-rbac-proxy/0.log" Dec 06 11:59:23 crc kubenswrapper[4678]: I1206 11:59:23.272068 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n2dhr_257b8a98-1ec9-4e1e-aef6-4f97cbe50546/frr/0.log" Dec 06 11:59:23 crc kubenswrapper[4678]: I1206 11:59:23.575138 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qjlkq_e6542baf-28b2-4844-991e-eb0f57d45231/speaker/0.log" Dec 06 11:59:38 crc kubenswrapper[4678]: I1206 11:59:38.123573 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/util/0.log" Dec 06 11:59:38 crc kubenswrapper[4678]: I1206 11:59:38.318128 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/pull/0.log" Dec 06 11:59:38 crc kubenswrapper[4678]: I1206 11:59:38.377217 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/pull/0.log" Dec 06 11:59:38 crc kubenswrapper[4678]: I1206 11:59:38.416588 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/util/0.log" Dec 06 11:59:38 crc kubenswrapper[4678]: I1206 11:59:38.531504 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/util/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.254300 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/pull/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.300220 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fndkz4_53e84367-27a1-4944-9e9b-bc9743e57398/extract/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.318064 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/util/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.606117 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/pull/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.631333 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/util/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.671228 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/pull/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.865781 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/pull/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.927932 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/extract/0.log" Dec 06 11:59:39 crc kubenswrapper[4678]: I1206 11:59:39.944155 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83fss8l_ad0dea4c-c200-4a62-a5df-e718bffb082e/util/0.log" Dec 06 11:59:40 crc kubenswrapper[4678]: I1206 11:59:40.110292 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-utilities/0.log" Dec 06 11:59:40 crc kubenswrapper[4678]: I1206 11:59:40.364076 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-utilities/0.log" Dec 06 11:59:40 crc kubenswrapper[4678]: I1206 11:59:40.384420 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-content/0.log" Dec 06 11:59:40 crc kubenswrapper[4678]: I1206 11:59:40.390215 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-content/0.log" Dec 06 11:59:40 crc kubenswrapper[4678]: I1206 11:59:40.610111 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-content/0.log" Dec 06 11:59:40 crc kubenswrapper[4678]: I1206 11:59:40.628580 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/extract-utilities/0.log" Dec 06 11:59:40 crc kubenswrapper[4678]: I1206 11:59:40.853577 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-utilities/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.156853 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-utilities/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.213792 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-content/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.249516 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-content/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.331621 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-49glg_62428199-c567-4609-8f46-2f31ae74136a/registry-server/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.413508 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-content/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.489233 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/extract-utilities/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.752000 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-s46l9_7dc2165b-f6c0-4386-b931-5e80f762dab5/marketplace-operator/0.log" Dec 06 11:59:41 crc kubenswrapper[4678]: I1206 11:59:41.810389 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-utilities/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.130623 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-content/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.146579 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-m2vrz_2c5e1af0-281c-42a7-bddc-0c7faa9b20e5/registry-server/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.187038 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-content/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.198178 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-utilities/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.398590 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-utilities/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.455212 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/extract-content/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.457929 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-utilities/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.617529 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-f64bv_18d2f32a-81c5-40e4-8912-6edffff17201/registry-server/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.716198 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-content/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.719875 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-content/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.739603 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-utilities/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.883229 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-utilities/0.log" Dec 06 11:59:42 crc kubenswrapper[4678]: I1206 11:59:42.915244 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/extract-content/0.log" Dec 06 11:59:43 crc kubenswrapper[4678]: I1206 11:59:43.447124 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-fc2vn_07116017-f2bf-4443-a09b-5df280727400/registry-server/0.log" Dec 06 11:59:59 crc kubenswrapper[4678]: I1206 11:59:59.505047 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 11:59:59 crc kubenswrapper[4678]: I1206 11:59:59.505585 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.146950 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f"] Dec 06 12:00:00 crc kubenswrapper[4678]: E1206 12:00:00.147714 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940f47d0-2037-441a-b80f-b94c1d638363" containerName="container-00" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.147738 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="940f47d0-2037-441a-b80f-b94c1d638363" containerName="container-00" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.148006 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="940f47d0-2037-441a-b80f-b94c1d638363" containerName="container-00" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.148746 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.150614 4678 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.151114 4678 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.163565 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f"] Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.267149 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm92t\" (UniqueName: \"kubernetes.io/projected/cef1388c-220c-4019-9a5b-f2ec7528e6b8-kube-api-access-qm92t\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.267210 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cef1388c-220c-4019-9a5b-f2ec7528e6b8-secret-volume\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.267247 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cef1388c-220c-4019-9a5b-f2ec7528e6b8-config-volume\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.368971 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm92t\" (UniqueName: \"kubernetes.io/projected/cef1388c-220c-4019-9a5b-f2ec7528e6b8-kube-api-access-qm92t\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.369031 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cef1388c-220c-4019-9a5b-f2ec7528e6b8-secret-volume\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.369070 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cef1388c-220c-4019-9a5b-f2ec7528e6b8-config-volume\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.369966 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cef1388c-220c-4019-9a5b-f2ec7528e6b8-config-volume\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.376709 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cef1388c-220c-4019-9a5b-f2ec7528e6b8-secret-volume\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.389976 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm92t\" (UniqueName: \"kubernetes.io/projected/cef1388c-220c-4019-9a5b-f2ec7528e6b8-kube-api-access-qm92t\") pod \"collect-profiles-29417040-6gf4f\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.470082 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:00 crc kubenswrapper[4678]: I1206 12:00:00.959271 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f"] Dec 06 12:00:01 crc kubenswrapper[4678]: I1206 12:00:01.934665 4678 generic.go:334] "Generic (PLEG): container finished" podID="cef1388c-220c-4019-9a5b-f2ec7528e6b8" containerID="5b27e5277724a208f0bc5851ed661029daec0059a92aa80faae5cc337bc3f520" exitCode=0 Dec 06 12:00:01 crc kubenswrapper[4678]: I1206 12:00:01.934854 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" event={"ID":"cef1388c-220c-4019-9a5b-f2ec7528e6b8","Type":"ContainerDied","Data":"5b27e5277724a208f0bc5851ed661029daec0059a92aa80faae5cc337bc3f520"} Dec 06 12:00:01 crc kubenswrapper[4678]: I1206 12:00:01.934920 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" event={"ID":"cef1388c-220c-4019-9a5b-f2ec7528e6b8","Type":"ContainerStarted","Data":"c08ef37020cbb02ff93eb933329cf43859644f13d3a87d72646ca68ff2abdb78"} Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.309087 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.334191 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm92t\" (UniqueName: \"kubernetes.io/projected/cef1388c-220c-4019-9a5b-f2ec7528e6b8-kube-api-access-qm92t\") pod \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.334378 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cef1388c-220c-4019-9a5b-f2ec7528e6b8-secret-volume\") pod \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.334462 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cef1388c-220c-4019-9a5b-f2ec7528e6b8-config-volume\") pod \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\" (UID: \"cef1388c-220c-4019-9a5b-f2ec7528e6b8\") " Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.335448 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cef1388c-220c-4019-9a5b-f2ec7528e6b8-config-volume" (OuterVolumeSpecName: "config-volume") pod "cef1388c-220c-4019-9a5b-f2ec7528e6b8" (UID: "cef1388c-220c-4019-9a5b-f2ec7528e6b8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.354797 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cef1388c-220c-4019-9a5b-f2ec7528e6b8-kube-api-access-qm92t" (OuterVolumeSpecName: "kube-api-access-qm92t") pod "cef1388c-220c-4019-9a5b-f2ec7528e6b8" (UID: "cef1388c-220c-4019-9a5b-f2ec7528e6b8"). InnerVolumeSpecName "kube-api-access-qm92t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.360960 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cef1388c-220c-4019-9a5b-f2ec7528e6b8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "cef1388c-220c-4019-9a5b-f2ec7528e6b8" (UID: "cef1388c-220c-4019-9a5b-f2ec7528e6b8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.436714 4678 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/cef1388c-220c-4019-9a5b-f2ec7528e6b8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.436765 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm92t\" (UniqueName: \"kubernetes.io/projected/cef1388c-220c-4019-9a5b-f2ec7528e6b8-kube-api-access-qm92t\") on node \"crc\" DevicePath \"\"" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.436781 4678 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/cef1388c-220c-4019-9a5b-f2ec7528e6b8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.952470 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" event={"ID":"cef1388c-220c-4019-9a5b-f2ec7528e6b8","Type":"ContainerDied","Data":"c08ef37020cbb02ff93eb933329cf43859644f13d3a87d72646ca68ff2abdb78"} Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.952521 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c08ef37020cbb02ff93eb933329cf43859644f13d3a87d72646ca68ff2abdb78" Dec 06 12:00:03 crc kubenswrapper[4678]: I1206 12:00:03.952552 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29417040-6gf4f" Dec 06 12:00:04 crc kubenswrapper[4678]: I1206 12:00:04.406460 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7"] Dec 06 12:00:04 crc kubenswrapper[4678]: I1206 12:00:04.420308 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29416995-c9nt7"] Dec 06 12:00:05 crc kubenswrapper[4678]: I1206 12:00:05.508894 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="182ea1b6-9b81-4b6b-ac80-7f6e0be788fa" path="/var/lib/kubelet/pods/182ea1b6-9b81-4b6b-ac80-7f6e0be788fa/volumes" Dec 06 12:00:29 crc kubenswrapper[4678]: I1206 12:00:29.504951 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 12:00:29 crc kubenswrapper[4678]: I1206 12:00:29.506403 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 12:00:49 crc kubenswrapper[4678]: I1206 12:00:49.342040 4678 scope.go:117] "RemoveContainer" containerID="00c4679b1e037722d88f9c27bc56eefd1a6863fca497487c2c7f860345bc243f" Dec 06 12:00:59 crc kubenswrapper[4678]: I1206 12:00:59.505574 4678 patch_prober.go:28] interesting pod/machine-config-daemon-bkhrd container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 06 12:00:59 crc kubenswrapper[4678]: I1206 12:00:59.506057 4678 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 06 12:00:59 crc kubenswrapper[4678]: I1206 12:00:59.506100 4678 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" Dec 06 12:00:59 crc kubenswrapper[4678]: I1206 12:00:59.506802 4678 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412"} pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 06 12:00:59 crc kubenswrapper[4678]: I1206 12:00:59.506852 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerName="machine-config-daemon" containerID="cri-o://a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" gracePeriod=600 Dec 06 12:00:59 crc kubenswrapper[4678]: E1206 12:00:59.638418 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.159089 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29417041-k4z7l"] Dec 06 12:01:00 crc kubenswrapper[4678]: E1206 12:01:00.159678 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cef1388c-220c-4019-9a5b-f2ec7528e6b8" containerName="collect-profiles" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.159706 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="cef1388c-220c-4019-9a5b-f2ec7528e6b8" containerName="collect-profiles" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.159979 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="cef1388c-220c-4019-9a5b-f2ec7528e6b8" containerName="collect-profiles" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.160730 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.172968 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29417041-k4z7l"] Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.278521 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-config-data\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.278777 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-combined-ca-bundle\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.278855 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-fernet-keys\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.278970 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6dsv\" (UniqueName: \"kubernetes.io/projected/8893dcb3-ddb0-49c5-b412-015d2854ac00-kube-api-access-r6dsv\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.380204 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-config-data\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.380266 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-combined-ca-bundle\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.380287 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-fernet-keys\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.380354 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6dsv\" (UniqueName: \"kubernetes.io/projected/8893dcb3-ddb0-49c5-b412-015d2854ac00-kube-api-access-r6dsv\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.388415 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-fernet-keys\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.392392 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-config-data\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.400075 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6dsv\" (UniqueName: \"kubernetes.io/projected/8893dcb3-ddb0-49c5-b412-015d2854ac00-kube-api-access-r6dsv\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.421627 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-combined-ca-bundle\") pod \"keystone-cron-29417041-k4z7l\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.462143 4678 generic.go:334] "Generic (PLEG): container finished" podID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" exitCode=0 Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.462182 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerDied","Data":"a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412"} Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.462215 4678 scope.go:117] "RemoveContainer" containerID="a8c3c8b241bd53b83426412a64cbd2cddf1cbe5ae5bf0d01ab9debe1412cf227" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.462596 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:01:00 crc kubenswrapper[4678]: E1206 12:01:00.462858 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:01:00 crc kubenswrapper[4678]: I1206 12:01:00.482470 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:01 crc kubenswrapper[4678]: I1206 12:01:01.029054 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29417041-k4z7l"] Dec 06 12:01:01 crc kubenswrapper[4678]: I1206 12:01:01.491588 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29417041-k4z7l" event={"ID":"8893dcb3-ddb0-49c5-b412-015d2854ac00","Type":"ContainerStarted","Data":"4383854664d67bcf83ba799c4c38b4eb75e6e71780b7467db9a9bd412b30f9f5"} Dec 06 12:01:01 crc kubenswrapper[4678]: I1206 12:01:01.491896 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29417041-k4z7l" event={"ID":"8893dcb3-ddb0-49c5-b412-015d2854ac00","Type":"ContainerStarted","Data":"a8182bb144b9a3c4e3948a434513787fca2ea2b24f87a215748f9c319288011a"} Dec 06 12:01:04 crc kubenswrapper[4678]: I1206 12:01:04.517505 4678 generic.go:334] "Generic (PLEG): container finished" podID="8893dcb3-ddb0-49c5-b412-015d2854ac00" containerID="4383854664d67bcf83ba799c4c38b4eb75e6e71780b7467db9a9bd412b30f9f5" exitCode=0 Dec 06 12:01:04 crc kubenswrapper[4678]: I1206 12:01:04.517565 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29417041-k4z7l" event={"ID":"8893dcb3-ddb0-49c5-b412-015d2854ac00","Type":"ContainerDied","Data":"4383854664d67bcf83ba799c4c38b4eb75e6e71780b7467db9a9bd412b30f9f5"} Dec 06 12:01:05 crc kubenswrapper[4678]: I1206 12:01:05.865699 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:05 crc kubenswrapper[4678]: I1206 12:01:05.995635 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6dsv\" (UniqueName: \"kubernetes.io/projected/8893dcb3-ddb0-49c5-b412-015d2854ac00-kube-api-access-r6dsv\") pod \"8893dcb3-ddb0-49c5-b412-015d2854ac00\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " Dec 06 12:01:05 crc kubenswrapper[4678]: I1206 12:01:05.995857 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-config-data\") pod \"8893dcb3-ddb0-49c5-b412-015d2854ac00\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " Dec 06 12:01:05 crc kubenswrapper[4678]: I1206 12:01:05.995919 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-fernet-keys\") pod \"8893dcb3-ddb0-49c5-b412-015d2854ac00\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:05.996738 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-combined-ca-bundle\") pod \"8893dcb3-ddb0-49c5-b412-015d2854ac00\" (UID: \"8893dcb3-ddb0-49c5-b412-015d2854ac00\") " Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.001194 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8893dcb3-ddb0-49c5-b412-015d2854ac00" (UID: "8893dcb3-ddb0-49c5-b412-015d2854ac00"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.005635 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8893dcb3-ddb0-49c5-b412-015d2854ac00-kube-api-access-r6dsv" (OuterVolumeSpecName: "kube-api-access-r6dsv") pod "8893dcb3-ddb0-49c5-b412-015d2854ac00" (UID: "8893dcb3-ddb0-49c5-b412-015d2854ac00"). InnerVolumeSpecName "kube-api-access-r6dsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.028861 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8893dcb3-ddb0-49c5-b412-015d2854ac00" (UID: "8893dcb3-ddb0-49c5-b412-015d2854ac00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.047665 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-config-data" (OuterVolumeSpecName: "config-data") pod "8893dcb3-ddb0-49c5-b412-015d2854ac00" (UID: "8893dcb3-ddb0-49c5-b412-015d2854ac00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.100163 4678 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.100205 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6dsv\" (UniqueName: \"kubernetes.io/projected/8893dcb3-ddb0-49c5-b412-015d2854ac00-kube-api-access-r6dsv\") on node \"crc\" DevicePath \"\"" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.100220 4678 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-config-data\") on node \"crc\" DevicePath \"\"" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.100231 4678 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8893dcb3-ddb0-49c5-b412-015d2854ac00-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.540044 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29417041-k4z7l" event={"ID":"8893dcb3-ddb0-49c5-b412-015d2854ac00","Type":"ContainerDied","Data":"a8182bb144b9a3c4e3948a434513787fca2ea2b24f87a215748f9c319288011a"} Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.540101 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8182bb144b9a3c4e3948a434513787fca2ea2b24f87a215748f9c319288011a" Dec 06 12:01:06 crc kubenswrapper[4678]: I1206 12:01:06.540141 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29417041-k4z7l" Dec 06 12:01:15 crc kubenswrapper[4678]: I1206 12:01:15.487969 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:01:15 crc kubenswrapper[4678]: E1206 12:01:15.488714 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:01:29 crc kubenswrapper[4678]: I1206 12:01:29.480845 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:01:29 crc kubenswrapper[4678]: E1206 12:01:29.481842 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.064190 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gpdxm"] Dec 06 12:01:34 crc kubenswrapper[4678]: E1206 12:01:34.065015 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8893dcb3-ddb0-49c5-b412-015d2854ac00" containerName="keystone-cron" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.065028 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="8893dcb3-ddb0-49c5-b412-015d2854ac00" containerName="keystone-cron" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.065226 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="8893dcb3-ddb0-49c5-b412-015d2854ac00" containerName="keystone-cron" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.066465 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.076945 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpdxm"] Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.225743 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9bgz\" (UniqueName: \"kubernetes.io/projected/792e4b72-b8c2-4831-aac9-9deaf4b709da-kube-api-access-m9bgz\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.225830 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-utilities\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.225971 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-catalog-content\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.328205 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-utilities\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.328368 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-catalog-content\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.328501 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9bgz\" (UniqueName: \"kubernetes.io/projected/792e4b72-b8c2-4831-aac9-9deaf4b709da-kube-api-access-m9bgz\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.328727 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-utilities\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.329269 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-catalog-content\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.355712 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9bgz\" (UniqueName: \"kubernetes.io/projected/792e4b72-b8c2-4831-aac9-9deaf4b709da-kube-api-access-m9bgz\") pod \"community-operators-gpdxm\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.399237 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:34 crc kubenswrapper[4678]: I1206 12:01:34.995368 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gpdxm"] Dec 06 12:01:34 crc kubenswrapper[4678]: W1206 12:01:34.997673 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod792e4b72_b8c2_4831_aac9_9deaf4b709da.slice/crio-e404770961bb9bd8b5b0b4edac59c10a771c878aa8e8be82a2caf121a225577c WatchSource:0}: Error finding container e404770961bb9bd8b5b0b4edac59c10a771c878aa8e8be82a2caf121a225577c: Status 404 returned error can't find the container with id e404770961bb9bd8b5b0b4edac59c10a771c878aa8e8be82a2caf121a225577c Dec 06 12:01:35 crc kubenswrapper[4678]: I1206 12:01:35.889679 4678 generic.go:334] "Generic (PLEG): container finished" podID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerID="e7de5f64e495a6541028078b1d924cda2e22529c1422ce945aa6d47a2161018d" exitCode=0 Dec 06 12:01:35 crc kubenswrapper[4678]: I1206 12:01:35.889948 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpdxm" event={"ID":"792e4b72-b8c2-4831-aac9-9deaf4b709da","Type":"ContainerDied","Data":"e7de5f64e495a6541028078b1d924cda2e22529c1422ce945aa6d47a2161018d"} Dec 06 12:01:35 crc kubenswrapper[4678]: I1206 12:01:35.890003 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpdxm" event={"ID":"792e4b72-b8c2-4831-aac9-9deaf4b709da","Type":"ContainerStarted","Data":"e404770961bb9bd8b5b0b4edac59c10a771c878aa8e8be82a2caf121a225577c"} Dec 06 12:01:35 crc kubenswrapper[4678]: I1206 12:01:35.891450 4678 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 06 12:01:37 crc kubenswrapper[4678]: I1206 12:01:37.915513 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpdxm" event={"ID":"792e4b72-b8c2-4831-aac9-9deaf4b709da","Type":"ContainerStarted","Data":"396d4162898985be1013388d4ce5cb6458f455a814a76229e76ac933c946ab31"} Dec 06 12:01:38 crc kubenswrapper[4678]: I1206 12:01:38.925046 4678 generic.go:334] "Generic (PLEG): container finished" podID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerID="396d4162898985be1013388d4ce5cb6458f455a814a76229e76ac933c946ab31" exitCode=0 Dec 06 12:01:38 crc kubenswrapper[4678]: I1206 12:01:38.925114 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpdxm" event={"ID":"792e4b72-b8c2-4831-aac9-9deaf4b709da","Type":"ContainerDied","Data":"396d4162898985be1013388d4ce5cb6458f455a814a76229e76ac933c946ab31"} Dec 06 12:01:39 crc kubenswrapper[4678]: I1206 12:01:39.934563 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpdxm" event={"ID":"792e4b72-b8c2-4831-aac9-9deaf4b709da","Type":"ContainerStarted","Data":"6e1a6de5eaf8eacab7cc247b7e412805980fafb4e5e2d8cb4fd9c2ccc1748f21"} Dec 06 12:01:39 crc kubenswrapper[4678]: I1206 12:01:39.957323 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gpdxm" podStartSLOduration=2.425228972 podStartE2EDuration="5.957305184s" podCreationTimestamp="2025-12-06 12:01:34 +0000 UTC" firstStartedPulling="2025-12-06 12:01:35.891273881 +0000 UTC m=+5100.734705320" lastFinishedPulling="2025-12-06 12:01:39.423350083 +0000 UTC m=+5104.266781532" observedRunningTime="2025-12-06 12:01:39.950275573 +0000 UTC m=+5104.793707012" watchObservedRunningTime="2025-12-06 12:01:39.957305184 +0000 UTC m=+5104.800736623" Dec 06 12:01:40 crc kubenswrapper[4678]: I1206 12:01:40.476301 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:01:40 crc kubenswrapper[4678]: E1206 12:01:40.476735 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:01:44 crc kubenswrapper[4678]: I1206 12:01:44.400219 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:44 crc kubenswrapper[4678]: I1206 12:01:44.400512 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:44 crc kubenswrapper[4678]: I1206 12:01:44.467527 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:45 crc kubenswrapper[4678]: I1206 12:01:45.025477 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:45 crc kubenswrapper[4678]: I1206 12:01:45.069625 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gpdxm"] Dec 06 12:01:47 crc kubenswrapper[4678]: I1206 12:01:47.033046 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gpdxm" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="registry-server" containerID="cri-o://6e1a6de5eaf8eacab7cc247b7e412805980fafb4e5e2d8cb4fd9c2ccc1748f21" gracePeriod=2 Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.056841 4678 generic.go:334] "Generic (PLEG): container finished" podID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerID="6e1a6de5eaf8eacab7cc247b7e412805980fafb4e5e2d8cb4fd9c2ccc1748f21" exitCode=0 Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.057299 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpdxm" event={"ID":"792e4b72-b8c2-4831-aac9-9deaf4b709da","Type":"ContainerDied","Data":"6e1a6de5eaf8eacab7cc247b7e412805980fafb4e5e2d8cb4fd9c2ccc1748f21"} Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.057351 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gpdxm" event={"ID":"792e4b72-b8c2-4831-aac9-9deaf4b709da","Type":"ContainerDied","Data":"e404770961bb9bd8b5b0b4edac59c10a771c878aa8e8be82a2caf121a225577c"} Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.057367 4678 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e404770961bb9bd8b5b0b4edac59c10a771c878aa8e8be82a2caf121a225577c" Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.110550 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.312895 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-utilities\") pod \"792e4b72-b8c2-4831-aac9-9deaf4b709da\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.313247 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9bgz\" (UniqueName: \"kubernetes.io/projected/792e4b72-b8c2-4831-aac9-9deaf4b709da-kube-api-access-m9bgz\") pod \"792e4b72-b8c2-4831-aac9-9deaf4b709da\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.313423 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-catalog-content\") pod \"792e4b72-b8c2-4831-aac9-9deaf4b709da\" (UID: \"792e4b72-b8c2-4831-aac9-9deaf4b709da\") " Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.313738 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-utilities" (OuterVolumeSpecName: "utilities") pod "792e4b72-b8c2-4831-aac9-9deaf4b709da" (UID: "792e4b72-b8c2-4831-aac9-9deaf4b709da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.314076 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.319727 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/792e4b72-b8c2-4831-aac9-9deaf4b709da-kube-api-access-m9bgz" (OuterVolumeSpecName: "kube-api-access-m9bgz") pod "792e4b72-b8c2-4831-aac9-9deaf4b709da" (UID: "792e4b72-b8c2-4831-aac9-9deaf4b709da"). InnerVolumeSpecName "kube-api-access-m9bgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.363956 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "792e4b72-b8c2-4831-aac9-9deaf4b709da" (UID: "792e4b72-b8c2-4831-aac9-9deaf4b709da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.415404 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/792e4b72-b8c2-4831-aac9-9deaf4b709da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 12:01:48 crc kubenswrapper[4678]: I1206 12:01:48.415435 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9bgz\" (UniqueName: \"kubernetes.io/projected/792e4b72-b8c2-4831-aac9-9deaf4b709da-kube-api-access-m9bgz\") on node \"crc\" DevicePath \"\"" Dec 06 12:01:49 crc kubenswrapper[4678]: I1206 12:01:49.067163 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gpdxm" Dec 06 12:01:49 crc kubenswrapper[4678]: I1206 12:01:49.107273 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gpdxm"] Dec 06 12:01:49 crc kubenswrapper[4678]: I1206 12:01:49.123475 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gpdxm"] Dec 06 12:01:49 crc kubenswrapper[4678]: I1206 12:01:49.485798 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" path="/var/lib/kubelet/pods/792e4b72-b8c2-4831-aac9-9deaf4b709da/volumes" Dec 06 12:01:52 crc kubenswrapper[4678]: I1206 12:01:52.476081 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:01:52 crc kubenswrapper[4678]: E1206 12:01:52.477036 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.476668 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:02:06 crc kubenswrapper[4678]: E1206 12:02:06.477566 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.863707 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rmmcn"] Dec 06 12:02:06 crc kubenswrapper[4678]: E1206 12:02:06.867012 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="extract-content" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.872092 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="extract-content" Dec 06 12:02:06 crc kubenswrapper[4678]: E1206 12:02:06.872259 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="extract-utilities" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.872379 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="extract-utilities" Dec 06 12:02:06 crc kubenswrapper[4678]: E1206 12:02:06.872572 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="registry-server" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.872709 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="registry-server" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.873280 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="792e4b72-b8c2-4831-aac9-9deaf4b709da" containerName="registry-server" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.880795 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmmcn"] Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.881056 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.953567 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-utilities\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.953824 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-catalog-content\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:06 crc kubenswrapper[4678]: I1206 12:02:06.954080 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2bq8\" (UniqueName: \"kubernetes.io/projected/661b47c5-2a9b-4cd7-9284-0e82efc69085-kube-api-access-q2bq8\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.054954 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-utilities\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.055034 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-catalog-content\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.055127 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2bq8\" (UniqueName: \"kubernetes.io/projected/661b47c5-2a9b-4cd7-9284-0e82efc69085-kube-api-access-q2bq8\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.055563 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-utilities\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.055683 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-catalog-content\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.229259 4678 generic.go:334] "Generic (PLEG): container finished" podID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerID="22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86" exitCode=0 Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.229300 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-shwqz/must-gather-2d8pq" event={"ID":"aebef180-e4b3-4f57-843c-24d89b64cb14","Type":"ContainerDied","Data":"22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86"} Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.229887 4678 scope.go:117] "RemoveContainer" containerID="22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.430305 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-shwqz_must-gather-2d8pq_aebef180-e4b3-4f57-843c-24d89b64cb14/gather/0.log" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.469313 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2bq8\" (UniqueName: \"kubernetes.io/projected/661b47c5-2a9b-4cd7-9284-0e82efc69085-kube-api-access-q2bq8\") pod \"redhat-marketplace-rmmcn\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.522367 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:07 crc kubenswrapper[4678]: I1206 12:02:07.998644 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmmcn"] Dec 06 12:02:08 crc kubenswrapper[4678]: I1206 12:02:08.239688 4678 generic.go:334] "Generic (PLEG): container finished" podID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerID="370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b" exitCode=0 Dec 06 12:02:08 crc kubenswrapper[4678]: I1206 12:02:08.239729 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmmcn" event={"ID":"661b47c5-2a9b-4cd7-9284-0e82efc69085","Type":"ContainerDied","Data":"370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b"} Dec 06 12:02:08 crc kubenswrapper[4678]: I1206 12:02:08.239919 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmmcn" event={"ID":"661b47c5-2a9b-4cd7-9284-0e82efc69085","Type":"ContainerStarted","Data":"46128d1066c6beb00533d42c3098a30a58df0a5b56db38c0c8e2a5f6c2c7b40d"} Dec 06 12:02:08 crc kubenswrapper[4678]: I1206 12:02:08.976984 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mkvgp"] Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.007667 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mkvgp"] Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.007799 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.198367 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-catalog-content\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.198420 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-utilities\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.198495 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64rhp\" (UniqueName: \"kubernetes.io/projected/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-kube-api-access-64rhp\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.299796 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64rhp\" (UniqueName: \"kubernetes.io/projected/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-kube-api-access-64rhp\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.300046 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-catalog-content\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.300141 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-utilities\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.300547 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-catalog-content\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.300602 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-utilities\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.341144 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64rhp\" (UniqueName: \"kubernetes.io/projected/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-kube-api-access-64rhp\") pod \"redhat-operators-mkvgp\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.359932 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:09 crc kubenswrapper[4678]: I1206 12:02:09.967717 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mkvgp"] Dec 06 12:02:09 crc kubenswrapper[4678]: E1206 12:02:09.970954 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod661b47c5_2a9b_4cd7_9284_0e82efc69085.slice/crio-b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd.scope\": RecentStats: unable to find data in memory cache]" Dec 06 12:02:09 crc kubenswrapper[4678]: W1206 12:02:09.976611 4678 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64cb1753_8a79_4bb9_91c2_639fac1eeb7e.slice/crio-ed9d5f836484204f26e48d7481528fc27c6553bea69be44e269609493da1fa39 WatchSource:0}: Error finding container ed9d5f836484204f26e48d7481528fc27c6553bea69be44e269609493da1fa39: Status 404 returned error can't find the container with id ed9d5f836484204f26e48d7481528fc27c6553bea69be44e269609493da1fa39 Dec 06 12:02:10 crc kubenswrapper[4678]: I1206 12:02:10.270843 4678 generic.go:334] "Generic (PLEG): container finished" podID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerID="b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd" exitCode=0 Dec 06 12:02:10 crc kubenswrapper[4678]: I1206 12:02:10.270927 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmmcn" event={"ID":"661b47c5-2a9b-4cd7-9284-0e82efc69085","Type":"ContainerDied","Data":"b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd"} Dec 06 12:02:10 crc kubenswrapper[4678]: I1206 12:02:10.274248 4678 generic.go:334] "Generic (PLEG): container finished" podID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerID="90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6" exitCode=0 Dec 06 12:02:10 crc kubenswrapper[4678]: I1206 12:02:10.274294 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkvgp" event={"ID":"64cb1753-8a79-4bb9-91c2-639fac1eeb7e","Type":"ContainerDied","Data":"90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6"} Dec 06 12:02:10 crc kubenswrapper[4678]: I1206 12:02:10.274320 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkvgp" event={"ID":"64cb1753-8a79-4bb9-91c2-639fac1eeb7e","Type":"ContainerStarted","Data":"ed9d5f836484204f26e48d7481528fc27c6553bea69be44e269609493da1fa39"} Dec 06 12:02:11 crc kubenswrapper[4678]: I1206 12:02:11.288238 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmmcn" event={"ID":"661b47c5-2a9b-4cd7-9284-0e82efc69085","Type":"ContainerStarted","Data":"0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8"} Dec 06 12:02:11 crc kubenswrapper[4678]: I1206 12:02:11.296849 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkvgp" event={"ID":"64cb1753-8a79-4bb9-91c2-639fac1eeb7e","Type":"ContainerStarted","Data":"8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96"} Dec 06 12:02:11 crc kubenswrapper[4678]: I1206 12:02:11.313727 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rmmcn" podStartSLOduration=2.9078248 podStartE2EDuration="5.313707997s" podCreationTimestamp="2025-12-06 12:02:06 +0000 UTC" firstStartedPulling="2025-12-06 12:02:08.241135754 +0000 UTC m=+5133.084567193" lastFinishedPulling="2025-12-06 12:02:10.647018951 +0000 UTC m=+5135.490450390" observedRunningTime="2025-12-06 12:02:11.304909511 +0000 UTC m=+5136.148340950" watchObservedRunningTime="2025-12-06 12:02:11.313707997 +0000 UTC m=+5136.157139436" Dec 06 12:02:14 crc kubenswrapper[4678]: I1206 12:02:14.329737 4678 generic.go:334] "Generic (PLEG): container finished" podID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerID="8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96" exitCode=0 Dec 06 12:02:14 crc kubenswrapper[4678]: I1206 12:02:14.329831 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkvgp" event={"ID":"64cb1753-8a79-4bb9-91c2-639fac1eeb7e","Type":"ContainerDied","Data":"8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96"} Dec 06 12:02:15 crc kubenswrapper[4678]: I1206 12:02:15.339982 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkvgp" event={"ID":"64cb1753-8a79-4bb9-91c2-639fac1eeb7e","Type":"ContainerStarted","Data":"55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9"} Dec 06 12:02:15 crc kubenswrapper[4678]: I1206 12:02:15.360792 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mkvgp" podStartSLOduration=2.909232692 podStartE2EDuration="7.360770726s" podCreationTimestamp="2025-12-06 12:02:08 +0000 UTC" firstStartedPulling="2025-12-06 12:02:10.275734765 +0000 UTC m=+5135.119166204" lastFinishedPulling="2025-12-06 12:02:14.727272799 +0000 UTC m=+5139.570704238" observedRunningTime="2025-12-06 12:02:15.356340137 +0000 UTC m=+5140.199771596" watchObservedRunningTime="2025-12-06 12:02:15.360770726 +0000 UTC m=+5140.204202175" Dec 06 12:02:17 crc kubenswrapper[4678]: I1206 12:02:17.523463 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:17 crc kubenswrapper[4678]: I1206 12:02:17.523788 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:17 crc kubenswrapper[4678]: I1206 12:02:17.617554 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:18 crc kubenswrapper[4678]: I1206 12:02:18.419045 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:18 crc kubenswrapper[4678]: I1206 12:02:18.554331 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmmcn"] Dec 06 12:02:19 crc kubenswrapper[4678]: I1206 12:02:19.360864 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:19 crc kubenswrapper[4678]: I1206 12:02:19.360908 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:20 crc kubenswrapper[4678]: I1206 12:02:20.381218 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rmmcn" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="registry-server" containerID="cri-o://0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8" gracePeriod=2 Dec 06 12:02:20 crc kubenswrapper[4678]: I1206 12:02:20.421973 4678 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mkvgp" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="registry-server" probeResult="failure" output=< Dec 06 12:02:20 crc kubenswrapper[4678]: timeout: failed to connect service ":50051" within 1s Dec 06 12:02:20 crc kubenswrapper[4678]: > Dec 06 12:02:20 crc kubenswrapper[4678]: I1206 12:02:20.486355 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:02:20 crc kubenswrapper[4678]: E1206 12:02:20.487285 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:02:20 crc kubenswrapper[4678]: I1206 12:02:20.929362 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.081144 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2bq8\" (UniqueName: \"kubernetes.io/projected/661b47c5-2a9b-4cd7-9284-0e82efc69085-kube-api-access-q2bq8\") pod \"661b47c5-2a9b-4cd7-9284-0e82efc69085\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.081197 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-utilities\") pod \"661b47c5-2a9b-4cd7-9284-0e82efc69085\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.081285 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-catalog-content\") pod \"661b47c5-2a9b-4cd7-9284-0e82efc69085\" (UID: \"661b47c5-2a9b-4cd7-9284-0e82efc69085\") " Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.088563 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-utilities" (OuterVolumeSpecName: "utilities") pod "661b47c5-2a9b-4cd7-9284-0e82efc69085" (UID: "661b47c5-2a9b-4cd7-9284-0e82efc69085"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.102538 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "661b47c5-2a9b-4cd7-9284-0e82efc69085" (UID: "661b47c5-2a9b-4cd7-9284-0e82efc69085"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.105777 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/661b47c5-2a9b-4cd7-9284-0e82efc69085-kube-api-access-q2bq8" (OuterVolumeSpecName: "kube-api-access-q2bq8") pod "661b47c5-2a9b-4cd7-9284-0e82efc69085" (UID: "661b47c5-2a9b-4cd7-9284-0e82efc69085"). InnerVolumeSpecName "kube-api-access-q2bq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.183456 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2bq8\" (UniqueName: \"kubernetes.io/projected/661b47c5-2a9b-4cd7-9284-0e82efc69085-kube-api-access-q2bq8\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.183506 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.183516 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/661b47c5-2a9b-4cd7-9284-0e82efc69085-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.392886 4678 generic.go:334] "Generic (PLEG): container finished" podID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerID="0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8" exitCode=0 Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.392944 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmmcn" event={"ID":"661b47c5-2a9b-4cd7-9284-0e82efc69085","Type":"ContainerDied","Data":"0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8"} Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.392951 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rmmcn" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.392983 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rmmcn" event={"ID":"661b47c5-2a9b-4cd7-9284-0e82efc69085","Type":"ContainerDied","Data":"46128d1066c6beb00533d42c3098a30a58df0a5b56db38c0c8e2a5f6c2c7b40d"} Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.393006 4678 scope.go:117] "RemoveContainer" containerID="0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.426177 4678 scope.go:117] "RemoveContainer" containerID="b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.436920 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmmcn"] Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.445809 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rmmcn"] Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.449693 4678 scope.go:117] "RemoveContainer" containerID="370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.490105 4678 scope.go:117] "RemoveContainer" containerID="0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8" Dec 06 12:02:21 crc kubenswrapper[4678]: E1206 12:02:21.490532 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8\": container with ID starting with 0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8 not found: ID does not exist" containerID="0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.490568 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8"} err="failed to get container status \"0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8\": rpc error: code = NotFound desc = could not find container \"0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8\": container with ID starting with 0bfecbd7535cf19704099b5bc8651b64f12daa81f60872d2b8b2ade95e11e0c8 not found: ID does not exist" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.490593 4678 scope.go:117] "RemoveContainer" containerID="b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.490743 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" path="/var/lib/kubelet/pods/661b47c5-2a9b-4cd7-9284-0e82efc69085/volumes" Dec 06 12:02:21 crc kubenswrapper[4678]: E1206 12:02:21.490856 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd\": container with ID starting with b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd not found: ID does not exist" containerID="b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.490882 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd"} err="failed to get container status \"b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd\": rpc error: code = NotFound desc = could not find container \"b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd\": container with ID starting with b09be26836d0370271940c599e1442ee2749a50f8eea22ae689a798f66df47bd not found: ID does not exist" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.490903 4678 scope.go:117] "RemoveContainer" containerID="370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b" Dec 06 12:02:21 crc kubenswrapper[4678]: E1206 12:02:21.491152 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b\": container with ID starting with 370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b not found: ID does not exist" containerID="370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b" Dec 06 12:02:21 crc kubenswrapper[4678]: I1206 12:02:21.491176 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b"} err="failed to get container status \"370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b\": rpc error: code = NotFound desc = could not find container \"370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b\": container with ID starting with 370a47a2cd3898cd6f640dad8ad96a0b902192b31e2534d096be1897c4f7e76b not found: ID does not exist" Dec 06 12:02:24 crc kubenswrapper[4678]: I1206 12:02:24.471670 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-shwqz/must-gather-2d8pq"] Dec 06 12:02:24 crc kubenswrapper[4678]: I1206 12:02:24.472445 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-shwqz/must-gather-2d8pq" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerName="copy" containerID="cri-o://499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df" gracePeriod=2 Dec 06 12:02:24 crc kubenswrapper[4678]: I1206 12:02:24.481974 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-shwqz/must-gather-2d8pq"] Dec 06 12:02:24 crc kubenswrapper[4678]: I1206 12:02:24.924894 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-shwqz_must-gather-2d8pq_aebef180-e4b3-4f57-843c-24d89b64cb14/copy/0.log" Dec 06 12:02:24 crc kubenswrapper[4678]: I1206 12:02:24.925635 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.058522 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aebef180-e4b3-4f57-843c-24d89b64cb14-must-gather-output\") pod \"aebef180-e4b3-4f57-843c-24d89b64cb14\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.058743 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8h6f\" (UniqueName: \"kubernetes.io/projected/aebef180-e4b3-4f57-843c-24d89b64cb14-kube-api-access-k8h6f\") pod \"aebef180-e4b3-4f57-843c-24d89b64cb14\" (UID: \"aebef180-e4b3-4f57-843c-24d89b64cb14\") " Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.064606 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aebef180-e4b3-4f57-843c-24d89b64cb14-kube-api-access-k8h6f" (OuterVolumeSpecName: "kube-api-access-k8h6f") pod "aebef180-e4b3-4f57-843c-24d89b64cb14" (UID: "aebef180-e4b3-4f57-843c-24d89b64cb14"). InnerVolumeSpecName "kube-api-access-k8h6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.161440 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8h6f\" (UniqueName: \"kubernetes.io/projected/aebef180-e4b3-4f57-843c-24d89b64cb14-kube-api-access-k8h6f\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.198525 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aebef180-e4b3-4f57-843c-24d89b64cb14-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "aebef180-e4b3-4f57-843c-24d89b64cb14" (UID: "aebef180-e4b3-4f57-843c-24d89b64cb14"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.262901 4678 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/aebef180-e4b3-4f57-843c-24d89b64cb14-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.439326 4678 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-shwqz_must-gather-2d8pq_aebef180-e4b3-4f57-843c-24d89b64cb14/copy/0.log" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.439805 4678 generic.go:334] "Generic (PLEG): container finished" podID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerID="499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df" exitCode=143 Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.439851 4678 scope.go:117] "RemoveContainer" containerID="499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.439964 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-shwqz/must-gather-2d8pq" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.467610 4678 scope.go:117] "RemoveContainer" containerID="22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.497789 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" path="/var/lib/kubelet/pods/aebef180-e4b3-4f57-843c-24d89b64cb14/volumes" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.506740 4678 scope.go:117] "RemoveContainer" containerID="499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df" Dec 06 12:02:25 crc kubenswrapper[4678]: E1206 12:02:25.507159 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df\": container with ID starting with 499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df not found: ID does not exist" containerID="499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.507190 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df"} err="failed to get container status \"499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df\": rpc error: code = NotFound desc = could not find container \"499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df\": container with ID starting with 499696f7d9398d84cc6092a6fe9535b37ee9335691f442209fc937da58f924df not found: ID does not exist" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.507210 4678 scope.go:117] "RemoveContainer" containerID="22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86" Dec 06 12:02:25 crc kubenswrapper[4678]: E1206 12:02:25.507529 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86\": container with ID starting with 22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86 not found: ID does not exist" containerID="22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86" Dec 06 12:02:25 crc kubenswrapper[4678]: I1206 12:02:25.507552 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86"} err="failed to get container status \"22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86\": rpc error: code = NotFound desc = could not find container \"22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86\": container with ID starting with 22c6f4c19dd9ecacc756d730ba6b76f89163de63ba645abf751db19841930d86 not found: ID does not exist" Dec 06 12:02:29 crc kubenswrapper[4678]: I1206 12:02:29.439508 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:29 crc kubenswrapper[4678]: I1206 12:02:29.493842 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:29 crc kubenswrapper[4678]: I1206 12:02:29.682835 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mkvgp"] Dec 06 12:02:30 crc kubenswrapper[4678]: I1206 12:02:30.503519 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mkvgp" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="registry-server" containerID="cri-o://55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9" gracePeriod=2 Dec 06 12:02:30 crc kubenswrapper[4678]: I1206 12:02:30.943577 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.072394 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-utilities\") pod \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.072747 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-catalog-content\") pod \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.072934 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64rhp\" (UniqueName: \"kubernetes.io/projected/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-kube-api-access-64rhp\") pod \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\" (UID: \"64cb1753-8a79-4bb9-91c2-639fac1eeb7e\") " Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.073251 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-utilities" (OuterVolumeSpecName: "utilities") pod "64cb1753-8a79-4bb9-91c2-639fac1eeb7e" (UID: "64cb1753-8a79-4bb9-91c2-639fac1eeb7e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.075887 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.121936 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-kube-api-access-64rhp" (OuterVolumeSpecName: "kube-api-access-64rhp") pod "64cb1753-8a79-4bb9-91c2-639fac1eeb7e" (UID: "64cb1753-8a79-4bb9-91c2-639fac1eeb7e"). InnerVolumeSpecName "kube-api-access-64rhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.180633 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64rhp\" (UniqueName: \"kubernetes.io/projected/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-kube-api-access-64rhp\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.253562 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64cb1753-8a79-4bb9-91c2-639fac1eeb7e" (UID: "64cb1753-8a79-4bb9-91c2-639fac1eeb7e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.282113 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64cb1753-8a79-4bb9-91c2-639fac1eeb7e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.476124 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:02:31 crc kubenswrapper[4678]: E1206 12:02:31.476451 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.514603 4678 generic.go:334] "Generic (PLEG): container finished" podID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerID="55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9" exitCode=0 Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.514657 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mkvgp" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.514669 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkvgp" event={"ID":"64cb1753-8a79-4bb9-91c2-639fac1eeb7e","Type":"ContainerDied","Data":"55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9"} Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.514760 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mkvgp" event={"ID":"64cb1753-8a79-4bb9-91c2-639fac1eeb7e","Type":"ContainerDied","Data":"ed9d5f836484204f26e48d7481528fc27c6553bea69be44e269609493da1fa39"} Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.514819 4678 scope.go:117] "RemoveContainer" containerID="55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.558223 4678 scope.go:117] "RemoveContainer" containerID="8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.559246 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mkvgp"] Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.570984 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mkvgp"] Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.587844 4678 scope.go:117] "RemoveContainer" containerID="90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.636054 4678 scope.go:117] "RemoveContainer" containerID="55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9" Dec 06 12:02:31 crc kubenswrapper[4678]: E1206 12:02:31.636820 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9\": container with ID starting with 55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9 not found: ID does not exist" containerID="55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.636950 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9"} err="failed to get container status \"55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9\": rpc error: code = NotFound desc = could not find container \"55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9\": container with ID starting with 55b3cb57a562e96bd7357cb0038dc279e3afbc803f49ac96d1082974a1270ab9 not found: ID does not exist" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.637053 4678 scope.go:117] "RemoveContainer" containerID="8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96" Dec 06 12:02:31 crc kubenswrapper[4678]: E1206 12:02:31.637723 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96\": container with ID starting with 8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96 not found: ID does not exist" containerID="8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.637777 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96"} err="failed to get container status \"8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96\": rpc error: code = NotFound desc = could not find container \"8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96\": container with ID starting with 8b458bd3c18ac042ed4a105cd7041141291fd212d20a13a07df523197fd32c96 not found: ID does not exist" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.637815 4678 scope.go:117] "RemoveContainer" containerID="90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6" Dec 06 12:02:31 crc kubenswrapper[4678]: E1206 12:02:31.638168 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6\": container with ID starting with 90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6 not found: ID does not exist" containerID="90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6" Dec 06 12:02:31 crc kubenswrapper[4678]: I1206 12:02:31.638209 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6"} err="failed to get container status \"90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6\": rpc error: code = NotFound desc = could not find container \"90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6\": container with ID starting with 90deeceec93c72df676c035c1343c8a3e03400fe569b19235e6bd792bebb83a6 not found: ID does not exist" Dec 06 12:02:33 crc kubenswrapper[4678]: I1206 12:02:33.492451 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" path="/var/lib/kubelet/pods/64cb1753-8a79-4bb9-91c2-639fac1eeb7e/volumes" Dec 06 12:02:44 crc kubenswrapper[4678]: I1206 12:02:44.476274 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:02:44 crc kubenswrapper[4678]: E1206 12:02:44.478633 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:02:58 crc kubenswrapper[4678]: I1206 12:02:58.476253 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:02:58 crc kubenswrapper[4678]: E1206 12:02:58.477408 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:03:13 crc kubenswrapper[4678]: I1206 12:03:13.476211 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:03:13 crc kubenswrapper[4678]: E1206 12:03:13.477298 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:03:26 crc kubenswrapper[4678]: I1206 12:03:26.475743 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:03:26 crc kubenswrapper[4678]: E1206 12:03:26.476600 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:03:37 crc kubenswrapper[4678]: I1206 12:03:37.476155 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:03:37 crc kubenswrapper[4678]: E1206 12:03:37.477448 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:03:48 crc kubenswrapper[4678]: I1206 12:03:48.476439 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:03:48 crc kubenswrapper[4678]: E1206 12:03:48.477796 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:04:00 crc kubenswrapper[4678]: I1206 12:04:00.476605 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:04:00 crc kubenswrapper[4678]: E1206 12:04:00.477595 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:04:13 crc kubenswrapper[4678]: I1206 12:04:13.476125 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:04:13 crc kubenswrapper[4678]: E1206 12:04:13.476955 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:04:25 crc kubenswrapper[4678]: I1206 12:04:25.483135 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:04:25 crc kubenswrapper[4678]: E1206 12:04:25.485956 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:04:40 crc kubenswrapper[4678]: I1206 12:04:40.476905 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:04:40 crc kubenswrapper[4678]: E1206 12:04:40.478333 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:04:54 crc kubenswrapper[4678]: I1206 12:04:54.475993 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:04:54 crc kubenswrapper[4678]: E1206 12:04:54.478207 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:05:06 crc kubenswrapper[4678]: I1206 12:05:06.693099 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:05:06 crc kubenswrapper[4678]: E1206 12:05:06.694553 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:05:19 crc kubenswrapper[4678]: I1206 12:05:19.475688 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:05:19 crc kubenswrapper[4678]: E1206 12:05:19.476452 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:05:30 crc kubenswrapper[4678]: I1206 12:05:30.476661 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:05:30 crc kubenswrapper[4678]: E1206 12:05:30.477557 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:05:44 crc kubenswrapper[4678]: I1206 12:05:44.475747 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:05:44 crc kubenswrapper[4678]: E1206 12:05:44.476964 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:05:56 crc kubenswrapper[4678]: I1206 12:05:56.476234 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:05:56 crc kubenswrapper[4678]: E1206 12:05:56.477263 4678 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-bkhrd_openshift-machine-config-operator(a6b30be2-60a6-4733-be36-3f70c6ca6b03)\"" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" podUID="a6b30be2-60a6-4733-be36-3f70c6ca6b03" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.943214 4678 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-656wl"] Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944403 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="registry-server" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944441 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="registry-server" Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944466 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="extract-content" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944477 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="extract-content" Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944517 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="extract-content" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944529 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="extract-content" Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944552 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="registry-server" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944562 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="registry-server" Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944588 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="extract-utilities" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944599 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="extract-utilities" Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944623 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerName="gather" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944635 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerName="gather" Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944676 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerName="copy" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944688 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerName="copy" Dec 06 12:06:01 crc kubenswrapper[4678]: E1206 12:06:01.944724 4678 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="extract-utilities" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.944735 4678 state_mem.go:107] "Deleted CPUSet assignment" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="extract-utilities" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.945009 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerName="copy" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.945045 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="aebef180-e4b3-4f57-843c-24d89b64cb14" containerName="gather" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.945077 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="661b47c5-2a9b-4cd7-9284-0e82efc69085" containerName="registry-server" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.945095 4678 memory_manager.go:354] "RemoveStaleState removing state" podUID="64cb1753-8a79-4bb9-91c2-639fac1eeb7e" containerName="registry-server" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.947215 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:01 crc kubenswrapper[4678]: I1206 12:06:01.990537 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-656wl"] Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.050898 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-utilities\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.051131 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjjtb\" (UniqueName: \"kubernetes.io/projected/09340311-88d3-44e1-9a8b-7124892d51c6-kube-api-access-kjjtb\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.051175 4678 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-catalog-content\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.153166 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-utilities\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.153360 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjjtb\" (UniqueName: \"kubernetes.io/projected/09340311-88d3-44e1-9a8b-7124892d51c6-kube-api-access-kjjtb\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.153422 4678 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-catalog-content\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.153561 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-utilities\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.153967 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-catalog-content\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.177176 4678 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjjtb\" (UniqueName: \"kubernetes.io/projected/09340311-88d3-44e1-9a8b-7124892d51c6-kube-api-access-kjjtb\") pod \"certified-operators-656wl\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.294598 4678 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:02 crc kubenswrapper[4678]: I1206 12:06:02.812640 4678 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-656wl"] Dec 06 12:06:03 crc kubenswrapper[4678]: I1206 12:06:03.040046 4678 generic.go:334] "Generic (PLEG): container finished" podID="09340311-88d3-44e1-9a8b-7124892d51c6" containerID="8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f" exitCode=0 Dec 06 12:06:03 crc kubenswrapper[4678]: I1206 12:06:03.041155 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-656wl" event={"ID":"09340311-88d3-44e1-9a8b-7124892d51c6","Type":"ContainerDied","Data":"8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f"} Dec 06 12:06:03 crc kubenswrapper[4678]: I1206 12:06:03.041267 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-656wl" event={"ID":"09340311-88d3-44e1-9a8b-7124892d51c6","Type":"ContainerStarted","Data":"b76ab5d5370db0548030c31748ffee9cbf84c065cb9931df32b3d06cfa897bd2"} Dec 06 12:06:04 crc kubenswrapper[4678]: I1206 12:06:04.052742 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-656wl" event={"ID":"09340311-88d3-44e1-9a8b-7124892d51c6","Type":"ContainerStarted","Data":"b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0"} Dec 06 12:06:05 crc kubenswrapper[4678]: I1206 12:06:05.065002 4678 generic.go:334] "Generic (PLEG): container finished" podID="09340311-88d3-44e1-9a8b-7124892d51c6" containerID="b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0" exitCode=0 Dec 06 12:06:05 crc kubenswrapper[4678]: I1206 12:06:05.065090 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-656wl" event={"ID":"09340311-88d3-44e1-9a8b-7124892d51c6","Type":"ContainerDied","Data":"b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0"} Dec 06 12:06:06 crc kubenswrapper[4678]: I1206 12:06:06.075587 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-656wl" event={"ID":"09340311-88d3-44e1-9a8b-7124892d51c6","Type":"ContainerStarted","Data":"0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af"} Dec 06 12:06:06 crc kubenswrapper[4678]: I1206 12:06:06.097323 4678 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-656wl" podStartSLOduration=2.414653561 podStartE2EDuration="5.097303315s" podCreationTimestamp="2025-12-06 12:06:01 +0000 UTC" firstStartedPulling="2025-12-06 12:06:03.041873078 +0000 UTC m=+5367.885304507" lastFinishedPulling="2025-12-06 12:06:05.724522822 +0000 UTC m=+5370.567954261" observedRunningTime="2025-12-06 12:06:06.090500712 +0000 UTC m=+5370.933932141" watchObservedRunningTime="2025-12-06 12:06:06.097303315 +0000 UTC m=+5370.940734754" Dec 06 12:06:11 crc kubenswrapper[4678]: I1206 12:06:11.478359 4678 scope.go:117] "RemoveContainer" containerID="a210aff15fc23b61617b494babab1db96e474c049ae95578f5efae3f98fa6412" Dec 06 12:06:12 crc kubenswrapper[4678]: I1206 12:06:12.135434 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-bkhrd" event={"ID":"a6b30be2-60a6-4733-be36-3f70c6ca6b03","Type":"ContainerStarted","Data":"0989c0bdbe3f2ec02fb00074a17dd19d1cad7b20a850c7c34c33280b04328986"} Dec 06 12:06:12 crc kubenswrapper[4678]: I1206 12:06:12.295211 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:12 crc kubenswrapper[4678]: I1206 12:06:12.295561 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:12 crc kubenswrapper[4678]: I1206 12:06:12.343990 4678 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:13 crc kubenswrapper[4678]: I1206 12:06:13.193767 4678 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:13 crc kubenswrapper[4678]: I1206 12:06:13.241446 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-656wl"] Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.161845 4678 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-656wl" podUID="09340311-88d3-44e1-9a8b-7124892d51c6" containerName="registry-server" containerID="cri-o://0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af" gracePeriod=2 Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.593990 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.724576 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-utilities" (OuterVolumeSpecName: "utilities") pod "09340311-88d3-44e1-9a8b-7124892d51c6" (UID: "09340311-88d3-44e1-9a8b-7124892d51c6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.725308 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-utilities\") pod \"09340311-88d3-44e1-9a8b-7124892d51c6\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.725376 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjjtb\" (UniqueName: \"kubernetes.io/projected/09340311-88d3-44e1-9a8b-7124892d51c6-kube-api-access-kjjtb\") pod \"09340311-88d3-44e1-9a8b-7124892d51c6\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.725474 4678 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-catalog-content\") pod \"09340311-88d3-44e1-9a8b-7124892d51c6\" (UID: \"09340311-88d3-44e1-9a8b-7124892d51c6\") " Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.728151 4678 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-utilities\") on node \"crc\" DevicePath \"\"" Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.733668 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09340311-88d3-44e1-9a8b-7124892d51c6-kube-api-access-kjjtb" (OuterVolumeSpecName: "kube-api-access-kjjtb") pod "09340311-88d3-44e1-9a8b-7124892d51c6" (UID: "09340311-88d3-44e1-9a8b-7124892d51c6"). InnerVolumeSpecName "kube-api-access-kjjtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.774218 4678 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09340311-88d3-44e1-9a8b-7124892d51c6" (UID: "09340311-88d3-44e1-9a8b-7124892d51c6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.850674 4678 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjjtb\" (UniqueName: \"kubernetes.io/projected/09340311-88d3-44e1-9a8b-7124892d51c6-kube-api-access-kjjtb\") on node \"crc\" DevicePath \"\"" Dec 06 12:06:15 crc kubenswrapper[4678]: I1206 12:06:15.850724 4678 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09340311-88d3-44e1-9a8b-7124892d51c6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.178291 4678 generic.go:334] "Generic (PLEG): container finished" podID="09340311-88d3-44e1-9a8b-7124892d51c6" containerID="0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af" exitCode=0 Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.178332 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-656wl" event={"ID":"09340311-88d3-44e1-9a8b-7124892d51c6","Type":"ContainerDied","Data":"0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af"} Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.178359 4678 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-656wl" event={"ID":"09340311-88d3-44e1-9a8b-7124892d51c6","Type":"ContainerDied","Data":"b76ab5d5370db0548030c31748ffee9cbf84c065cb9931df32b3d06cfa897bd2"} Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.178380 4678 scope.go:117] "RemoveContainer" containerID="0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.178442 4678 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-656wl" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.201238 4678 scope.go:117] "RemoveContainer" containerID="b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.237636 4678 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-656wl"] Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.254705 4678 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-656wl"] Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.255389 4678 scope.go:117] "RemoveContainer" containerID="8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.307892 4678 scope.go:117] "RemoveContainer" containerID="0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af" Dec 06 12:06:16 crc kubenswrapper[4678]: E1206 12:06:16.308318 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af\": container with ID starting with 0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af not found: ID does not exist" containerID="0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.308428 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af"} err="failed to get container status \"0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af\": rpc error: code = NotFound desc = could not find container \"0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af\": container with ID starting with 0ddbe9d5c91c50847b310d6e6df73f960a05a6ba65584bf0b37ad0a8e34c00af not found: ID does not exist" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.308546 4678 scope.go:117] "RemoveContainer" containerID="b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0" Dec 06 12:06:16 crc kubenswrapper[4678]: E1206 12:06:16.308920 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0\": container with ID starting with b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0 not found: ID does not exist" containerID="b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.308972 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0"} err="failed to get container status \"b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0\": rpc error: code = NotFound desc = could not find container \"b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0\": container with ID starting with b90ec9746c1917a6a6a0d2d8377382aea545c9d4db5424804b7b19fee94b9bc0 not found: ID does not exist" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.309008 4678 scope.go:117] "RemoveContainer" containerID="8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f" Dec 06 12:06:16 crc kubenswrapper[4678]: E1206 12:06:16.309309 4678 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f\": container with ID starting with 8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f not found: ID does not exist" containerID="8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f" Dec 06 12:06:16 crc kubenswrapper[4678]: I1206 12:06:16.309339 4678 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f"} err="failed to get container status \"8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f\": rpc error: code = NotFound desc = could not find container \"8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f\": container with ID starting with 8ec3334c4518b5b59610a6359806de9ab5bcf797fcb103e5c9e481626fe0118f not found: ID does not exist" Dec 06 12:06:16 crc kubenswrapper[4678]: E1206 12:06:16.406868 4678 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09340311_88d3_44e1_9a8b_7124892d51c6.slice\": RecentStats: unable to find data in memory cache]" Dec 06 12:06:17 crc kubenswrapper[4678]: I1206 12:06:17.491110 4678 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09340311-88d3-44e1-9a8b-7124892d51c6" path="/var/lib/kubelet/pods/09340311-88d3-44e1-9a8b-7124892d51c6/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515115016106024441 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015115016106017356 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015115003052016475 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015115003052015445 5ustar corecore